var/home/core/zuul-output/0000755000175000017500000000000015134264461014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134272267015502 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000221767415134272132020271 0ustar corecoreZtqikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD )?KYꋴI_翪|mvşo#oVݏKf+ovpZj!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!frH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{BpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&Va%ĉUHSR0=>u)oQCC|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ"0(HKkD4<80: M:'֥P!r "Lӓݰ@ 9n# " $fGgKQӦ4}Gn\^=-Y5PI dPN6 Ozځ/פ|5) F[ڣ$2*%&h v%9HN H~Q+oi?&۳)-nqK?2ސv/3,9ҮT9Cef˝49i.2DxatC<8iR/ƬйR֌vN8J"iJ. T>)qaY4ͬlyg "]BvW#99`TegõII kюHLa^c&/H^FFIu`2a$mc Ry+R:LڕDܓ>Y:]t.+|PT6=qWe0NƏw<6o3mv8k vGOfpEOkÈWȤMف lOc;SR&.w,qk>MPs+Xh4iyuGRd֞q鮺]m S{}]U kV0/ŜxtADx"Xh4|;XSxߵă@pE:y]/"(MCG`ʶϊGi+39#gNZYE:Qw9muB`9`LDhs4Ǩ9S`EkM{zB<˙ik; JD;;3!4 2Y.$Dwiu|+lO:k$]ԜYLUҞ6EmH>azʳ/A+ԀZk"f`.,ל{=wh|_qYj5M{K$gv>cDp"'0޽5xCNQ1G2})*'>fC۝'*)"5.E2IeD 2.ZdrN6Uœ=n8D-9޵JKw5ُJ,􋃓ZUꋼ0b1f87GՂ 1t_o}{Mr7KO0Ao-Y*Is\S:JzA(:i!eҎ\,f+,Ąt78~ڋ~?[F^.A'!,iGow3{'YToҝf5ޓ[he>=7S8DGZ@-#]f:Tm?L{F-8G#%.fM8Y='gیl0HڜHLK'Cw#)krWIk<1څ 9abHl:b3LjOq͂Ӥ=u8#E2;|z꽐vɀi^lUt␚ɓW%OVc8|*yI0U=nFGA`IC8p+C:!}Nh,mn>_MGiq'N~|z`|mu}r:"KiyGҪ$& hw#4qn?ܶХfm_Ov^ܶ[6j3ZN9t9ZMMM)I[Rχ/C|W䳮yI3MڼH9iEG&V 'x`u.̀ab7V<*EzfH{]:*6M x-v쳎M'.hO3p-IGh ܆hR ]zi2hB9'S_;I/d0oIU:m/~[*K1QA="D:V&f:{7N>^uU` c/X)mS5KC߄":{H)"%,!3w{"ZWÂk>/F?RJ>FIY*%5Hg}3Ď89؟N/pgÞ tJXB-Gjsٶ 3Gzp؍H|*cyp@\첹,[up`uV,\KCB\qGiW痃[?i?S{eϻl71X:݌>EEly(*SHN:ӫOq{{L$?Q{϶(F_Ej>3mqfΤP-j)H˧&8?a?2xĐ+EV؍x0bv6 fd1^ 2ӎԥ sZR cgu/bn/34'h9Dݥ:U:vV[ 'Mȥ@ەX㧿-p0?Q6 y2XN2_h~Cֆ֙82)=Ȓ7D- V)T? O/VFeUk'7KIT, WeՔ}-66V؅ʹ;T$pZ#@L; ?0]"2v[hׂ'cJ6H4bs+3(@z$.K!#Šj2ݢxK-di +9Hᇷ絻+ O.i2.I+69EVyw8//|~<ëng)P<xͯ~? fp,CǴ_BjDN^5)s('cBh+6ez0)_~zJz"ё`Z&Z![0rGBK 5G~<:H~W>;ٍVnSt%_!BZMMeccBҎÒJH+"ūyR}X~juPp- j\hЪQxchKaS,xS"cV8i8'-sOKB<չw"|{/MC8&%Og3E#O%`N)p#4YUh^ ɨڻ#Ch@(R &Z+<3ݰb/St=&yo|BL,1+t C<ˉvRfQ*e"T:*Dᰤ*~IClz^F6!ܠqK3%$E)~?wy,u'u() C>Gn} t]2_}!1NodI_Bǂ/^8\3m!'(Ֆ5Q&xo 8;'Jbo&XL_ʣ^^"Lq2E3,v1ɢu^}G7Z/qC^'+HDy=\]?d|9i,p?߼=\Ce"|Rݷ Q+=zxB.^Bld.HSntºB4~4]%.i|҂"? ~#ݤ[tfv3Ytck0O ͧ gP\|bЯ݃5H+޹na4p9/B@Dvܫs;/f֚Znϻ-|X!lk҃=pnUגZ6p| G;;74^l{Pclwů Հ}xcSu)6fbM/R(*ȴd.^Qw %"=nluOeH=t) Hİd/D!-Ɩ:;v8`vU~Ʉ!hX #'$2j1ܒZ˜bK@*`*#QA 9WykGk,8}B6{/) ݆Y~ 1;;|,ۇ=sxy+@{l/*+E2}`pNU`ZS̯窜qN8V ['4d!FmaX-6 y:1V(!L7,RPEd;)QϢ +RlWDžuF7LFֆoM~ar*EtIbW>jqour?qzJJaQ#-n`/$fhnqgTĔO5 ꐌSYXzv9[ezksA`<dkON৯s|&*pNaJه5B5H:W2% `6MRR'xZtfC$1aH_dx$1'/v^ZZ4`9);q`F"d1v>ժbLGd~MP%m x52LMF9 E"A,S Vo}\"X.2< 5FB΢u.`aJ#Tk’"D#cuCXȉ4 ՖK(KP|dZ1&8{9rLnMRф%V Ng2K|`ot.GSGd oE'!B'Nb1{8LW^9KbN;sö!`0ݘ/l+1L#B8U֕&*?V6N{դ}Y(INBKhx2 *MOenT.a~.E jG)j{=u^K+Ȫcv/w#MivX :)ǪCZUnAS`SK6OSxa3 W; K>窜̀'n 3u0?K@BS %fee}i]>̤+*l:\歶!IZ5>H;0)N.w7ߍ|+qUߤ^oå~4en\.cY[s'wSSۘf ?.D s}Y~/J[}jX^ޗ_-/̍ݥ*n./cus}]\>\\^'W_nAqC_oO-S_sOq?B}mmK2/@DJt}=xL@5MG0ZY,\S Eb uw:YɊ|ZԘ8'ˠ*>q/E b\ R%.aS qY>W Rlz!>Z.|<VD h5^6eM>y̆@ x>Lh!*<-lo_V684A飑i2#@+j3l૎S1@:G|gRcƈ?H(m>LC,HI~'.Op% ' c*Dp*cj|>z G` |]e*:nq!`{ qBAgPSO}E`́JPu#]' 3N+;fwt[wL X1!;W$*죓Ha-s>Vzk[~S_vD.yΕ`h9U|A܌ЃECTC Tnpצho!=V qy)U cigs^>sgv"4N9W_iI NRCǔd X1Lb.u@`X]nl}!:ViI[/SE un޷(ȊD0M^`MDN74Т C>F-}$A:XBgJWq&4ۓflq6TX)ى?Nwg>]dt*?Ű~{N_w7p682~ =WBX"XA:#u-9`x 92$4_>9WvTIj`+C2"s%DƖ|2H\2+AaTaBˮ}L@dr_Wfc>IdA Od[jlec=XJ|&+-T1m8NP$%s,ig\Z:h Ћ߉n!r}_\ \5 6 d#=&X^-kOwĝJO\Vj; )!eoB4F\jtctUb.L[3M8V|&jZz/@7aV),A[5TpUZL_?CU0E [%W%vl x٘3܎y,< )i7 Ո: tC`\?c%v7\Ct!$9iç$><+c~݊lz1H[E'2/clQ.I`AWOlw&5fH n`gMytdx)lwAK~GgbJI-tq5/i ?WǠr^C/1NEU<=co(k0Q~wˌ\g,\ rf\PUH,L#L7E"`0dq@zn~+CX|,l_B'9Dcuu|~z+G q|-bb^HcUha9ce1P[;qsA.Ǎ-]W‹y?ڕ^Pm:>I+Ȧ6' ,}U=̀*Eg.6_~OJ/8V ?ç&+|t><,BLqL򱷬dS{X6"X#-^䀕#{К4i̎'QIc(<ǩJi lc*n;YKO?md+F/[܋?TUXS-*1l=i6v=/DS&'i! I 鉊60ΓjB. `0 vf2SRmlNI S7t un/g{m|r4[^W&CҘ ?yq}{]\&Fŕs=h~Y̵M_, 5,ݲC4B_[UcȊgE5qT -W;4]9S_{Y,8?*%ĈAM`=uOZ?Ip^ױE\2uW3^ngkn(4]"{;Mmb܇ܵ,5C| fAפ PC,`:bx8Wno[#yDa4Mx>/RwBr7k?XS5Jծ4JTJ7Ǻ쵥ͣuՠ,(c8{7+C4ݶl%BOں{}z8~dkji/|'_*t_^r1"_(> ˏ \K.LwP!e28 k)ʊ54ȳy"HGȁ)^? (V1*gYe:Q\@a\(c-)[7RѬ_N*^^f`-~KbKoNFaVXQbZϖJdMJU+A]4a Q Igmw/Y{jƣ?8F" &PB> swކ?JүT .(G@/(G sb<:k~~ 6 ox R,yx%'=.G_ !־ol#GSJ, N;AD,2hDVy)bRxZ9؞?sZ. GmŠ#ũ(>-?Ө^,E=:8)D /9 VEj]mFl~]:St1i㸬vtSX8˪f$ L1/Ĺx119B%ԝ|NOڍv-h(v_ٗ63Ii4S//.hv_?/;-kS.IPݗՊEV)dV4(P2Ϊsz*ߧZ͈2wFtQDs]B Gb>8i, M)OVO)O"]tҞ5sLWQ@K?&+?FsXi+hӚǗ]@QwW DqRigŗo4o\߀q A:mw_.sۍ= ҍA]F&ո_DO_*Qt \hw+{VW(cR$$W1 J녜gi UGԞ!(kcoO1gMTV-h`e­1N4wwΣ[\QŘW&&G'x|Y{J請o箙*i\prg2&]޵}P2jp'EsAX 2o7 ԅ/Hpt곞F|ƀ x>ì!gq[1mY GҮ`+0/4(5ttqKlZFPi[ik *A]<|vzB@ȞDi*z>%xk "Hw>4:8 zPեkҕ!~ !YGގuw@fǟRFJ<xԯb W]&f"RGp;)yw`%iC%W8J$Su ZʍfYv}x$`vs(̀x%jICb09.Q( gG _4"q97_~[IiNѩp,xYV)Q>]ٱU0Q=LW.axR )\b2R\ gy N\B U \KүQ0%,~5WV( jxL*Iߎ8^DK3AyI ֆ1`ZK^OiDM'_}YQGq3yȖ3dП,Δjvt8&%8٠ʂ]FUvC%?Z=݋EM=G=ELu- vR-r0]kw&42-q,R[6%#jXgI2-jU%_0Bך&jmʳ&9> sGB,64՝fWpCݖm\{TYWLv*#0ٺ/`\frXnxX0C-~3-)`1"hs AWgܒr<ˎbd抰]L&}k׊| \1dH$\J0 AdZa\UF<).ɊzKm I]SS\m,g ~ژWҶ /^(XoJ=ԋ.Tg!ÃeFXv#w3z~Q@v<*X"iYKlЋ1eݣXg2]҆ta)sp0m+3bqgҌJe3\FLQvluS㚲0!y}jF3{ysӚ4l2~,IUewe5 ao&3==,-A$ J+9TWN.o(bчT-&,l2fiNy Ȼ$_^9Z%iuI ~Z $Dud>mҰnw7C;Ĝ=L`=X4[1bA/`Ƌi`(ˏOӬ/:U G Wm{܃%Lqf01mս2k_"00ܯ߹ڏA2կRzI@0ɯُ6`ph!i+pm+"gNE(LT\~{dB+PCxzX}! `t}f!$gOƣBj(rn92{ N)O"cv$G4ʬ^hSS PDh/7g{1MS] sz*x5#ncצd h/Xudswm{2FcO%`$}Ol\,a1mtt.NVMz.܏d0"j8ycO)Ph =1_|þǟ{?&{dhFA[!ݗI#deUpw/\w ܗI#LYYLMj0̻Sv6bE "0E' K ?{W6 Uojw]8Ǹ&N|l$HDc] dˊl(53I$uF4Nk30)o]7Y^>Cq@ OK?WJᆂe:٩ 'N5?o+L>o'uUop>W=NoiN0TTmy8XUOU\~"#xJ&)3L VuZ1ى *gKQpRjq~t>E@u9آS<eU//([T'U6x-yI۳iӞItO6.uo,<_~{)l \ut]4Bd~70 '$p[u :v!o4 t-!&$BiJDI6u"]>TΏh@^KKkv0 8Ol'1x;ݦeĽ ;yATk<[ּC S|YzKNmo2$ޢem0NۍxB, 3•mZ0[N`ZBP\hS! m5Nx-#=Bq Pk a@N;|=v J1xM+gP; ؙ$2lo!y7#E4@8XY"OzkDI%a83wIywFL3Y i@'9JO?;2 |Lu\G٠_O?gn"9M[=z9wC jT'r#^yچjj_+&Fu/%+4SpkT zJbEj׷5׽oWNAlumv#Ev@i{: Rl"ehRBY zxKCF\LƅIL5# 7a$Zddg#l1,7`I DYDaBYP,/h]Ըg̏ip=q |r#.سo`C7WbjI g\fWjiL+"If&>mLHtS-8+qt?F}úቿ"8'46,f0O%&"" iX %ƪZۙgԵ>k<\xHx[ѧ D~t8ѽAxZUKK$W_c<O5v$]<+_ǫj ,b43HgHA@y芁ϽODBNBK` I{з2 CXİku$ң Y8t8]'Hy7b&N``D[hWO}TW4<^|s8Ȥ0h'm9н׸:ygьcQ+ 1F@o %-e=r)pi Ic#kDDEc܃drm`}Fx6Kɂ-X],ܓ ir%9$`uxĈIV7(g~y4#."^;="]pɚ#ڸpEa˦P0j/aO:MZGBQeQ6\b}e4pu">(x%N-K0ZitD4ҳ+*XI J>4 r($ &m/?aou tPtsB$nB.)}F*}nҍanԳڗ;weԲ e^F*{&lB2|BʟI(߀PgPgB u^FLB u u7 ݜPe$݀PoPoB ^FLB  7 ߜPe$߀``B ^FhLB  7 4ܜe$4\A.60 PI~3!ҕ[uGgy[\I~t8zL\ge^tޕ&]I.s4*BJ'C"I,E\ 8L;sH<1D7n:=0=ޅg,&Pϟ@k$`Zd V2NC^耲\Q]aP4A8 IR07b>F*@ђg "i'}|l\@, 3`0! jeASwϡ/d~|p{4O3_|4zHsh+1E< ]({cy\05'leW[ޒ#xͰeM!_XuNU<ǣ$_Xҩ1C%~hBo8p=ˁcel }h04+)bd꿗qD%EP&>u,ڈB0EO9dJldSׇ3L"X/;K5OcW  0oLՏ*H'i:&fŢ<=]6`esk &/7 \8xs*3<؇ZyzVIȳB>UU j\H9H<*Rm:TZRTKLp8wY/KjBU(zJa5c}P!.3*biϺ1Ee¯I?!UN2蠘MjUYl~Wq2mJ?O /EԬj&nHRB?4j!.5 LB2|*Y0̉lf ~ ZGZuhq rWwE9S%W@pVsQgcql2Bkղ[?bТҕ-2F&u ʾn~u)R46F/Oh{m1 *yJ kk<\(:89~rC~B/׭ GpB[[t84|NdըV("F7@i`k[jXr9]˃ab'& #mvښ,S{/V+͛,~==j[,a3}Dr lrIPIm+(MTdɈ=h{84Em6ra̾[؟?Nl;߈ы"ى$ Q6H\Y#?!Q? eb~0 $in"pooeyU ՘ZyN^Ix<_Gί oJ.h[U'g>i:Ɲ'ܧ )ܡ Ku%|6zF᏷Z`2Rӳ`j\VjaEsB:8,P/v,FZ=>%8I mo&$I!+r1*ǏoIpz?TEST Ӵxo(bޓvˁz.v-9Rmt#uܛEjyb`ۊ> Af54$f*ْ$^$@gB= }ڃIr2)AI+4) =ٹX,Tmbs™NU5H:s_Xp]@@< cлsy0דd"GLK6Y_1I(b8b~\[ 0@drBDVj IP YPaֵS<~<%2^U|r\|۹vTˑÚN:cX섷1Jñ U"%?$-hb0L JtϦ+)jH<.a`?>Y +J*)Bu#'!heD)M,8ƹ5&D /`jLy/JSD"I7ռrd2& jQ뢑YE*b*# ,qdP?F3(`U;ds͔\E;*ұw٢x\6EK.wa0Qˣ5Q@yhf5)&UE'Ig3Vq'7~8]!\o "ZMUtIg8_{,8pUwC]٠٤-c*׽Ջ^1i\rD#1kx [ЛİZj$15{~nݚs z)]/*R7X5(b`<QfA.νѲze5IJv)Cp\-FԓΟ_W)[iGܕqUQɉI5֪To1BCbb24)><5zrd7f/@hBQ*^"^Zlp]>8G. ՘rM>XIQG-j4%K щ Agg# loXth.BrNꨋAayț~L>U!+U)@PuR0*n>hHb,-㶉Jt<)77p>Ƃg|?pѠ/hbDWS^)Et@],5z2&Y`t; 0lp. V \|)"EE CØ01S50Kcc)"EҜ¡ &bz 5gIG9)K9ZUbݲ+O?}Y bh*->e'svb>WMD*n{EkvEdClH:q[^;KKO$to9O8r h&W [F7 ?5,kЍ o-J=.Kݱ itfrKբ6Pe01)HٌAT c*3AW pX3eWU_lRgXw'b/񽯋"wƄNEh bYEH-K)W-1FR vVL5i#h&E|umrz*A}maQ /]8SWH2yŁjD J^=Oo1;AVk΂T&ptfJNc.w, tan."}%ǬM3g"xk%NP$Uft\+9 ɏHhgW`ܾчWU`K#d\NGudrr5WS ̠̥th>vbS#Z)ά/兹{Rqm0(S堐kyUCc SzSXJ~rF7ym;f))rX[G N`R-&EVZ9uEҿںgMaš6NsmE+aͺ49wƳਾ.ņLYjYUkj$+k3Rİ;`UucGi5dgl]&2fx 9ʹ'=[y|^@1_n)ί%a]/%\!S"* Qǯ >ԶRPـ M}RKlS3KmZ HԸٓ9FKBp|ƏkÖ`jGGʚ8?lG#&/#j&q,;FH\.YV =#h O :0/G;8dgΘCƠŴ5.3os,8zuUիHC&Oi7rrsi~P RFN?9iuؠA;+]("+M)`QnMw~]F?I\-*(i\zNuӶS9{gcCMGpb< ĝI&Ћ/o*EG;r7қE>l=[ςmt4U!c,Ü&69)1̝Fi|9l1yدyW<* c'U g;w8;5g#P }Y@0m.ƪĢ>E?O; =݈\Zz}9r~vIwc_3 꺶}TVmamb&uvd*1L'G4.@]xb\xp=T!M`²茾j Ճ$ƐkqUg߅PXp̈́~!;:܃ f xWkODYaqջo-$\YQ6Q? ¾?`YtRy]'E$Mם6MPTI Wkk :D&T"P4 %E)^$Ivh9l˞G'] rBW,)^ yX/%x0NI"ΐ8dҗ74W- @L bU4c Pe$d#hFc&uS$v} Λ8dޕO午p'/^+nAM;k/7/NVL^v)GwLڊ4&4L7ƪ낺"/BJaHa c Ԓh߰UD)V6[Uݙ7Q.LV2iN"ǧ7 OZHgS Z)}sMɌfǛS]q^_VwW=?n7czsϫӶ\Yj.0Ih$т>q";@k/: ;(矾=*>dmK_o]!m"hMUL05FVJ#C &|Q"cn֛.M&lC<+3t0{ѢB@ ÃLD"UR՗5Qȹ(JdQM< .Nå{QL.쌏]kL1Z1H8BtzͰ(((2i߶\,Aɛ}/:I!>u c,i 5 :BOW GIoy9 R4gqdХŖJe2_ jI p] ) i!j:t~Iig~ẑZgE <(x) &`Rav=Sw$-*l҆PJSF*-oM;scJm==' Uo)Ԫ4F!"CNY5"!P$q? GwX̳{gy/ t[^Z_zq#1mݰTE29Htr_9H;|/zVGҚ{ 4٣jq_Wi-[ll,C'Wl.˵l3Vħ5;˝U׵LUI~0Ôw 1<t)Ò@ =? u'ƔTnWѾtq 9uH*Iz)q9!Eҙ.*יkQ'4hNTAM4jD)N__ٻƍ,WMf[v hfit80ȒGq{,ʒ[)ɤ %ԭsuY|4a*.>2mHפ1B fl6i[<Y쭈8pqW7ugkyvtwSp?Qx[֥y8ea/8R* So cY7&6K[ϩ{@\%iqKuJ˫xC&m4MݲmTcFAH|8Ep>D,d*OS|6i;j?wFG's*_{lW>f}?==|Q/hw;] \ o\R~KbY%v>OI3/*] _ćW0t@C:1e[5P#S<@ߝMqB|p0j \=r_fg#c2T߳uJ}s{3LBxT8o>&Pd~fN:Z\/_33?"_c^\\b83O,GZ9,23pE?J~pO:nY"^E䂽4~K ՝djzi|A)Nazq9_%ӣ`E-SI! 3sNs q=[YcguMxlj4drS?tmkLvg+oYɕsw?F7ޛ T .xI='&߶x>|XX%BUΕ[s]0+ T^ P|\>۾^#޹oq 8=IgCE~h->hc;>]g7`@ $ۛ#LG^t*Dzo,is*ipebKޕUlcٌ p7B|dWV Ѩ[i|%RIyPccvWw`h'u9 Kxr9uch'Q w=ׅ~G/SPV5В+2f:V&g^Y~$+c%#0 8%"oY 翗p/?}m줬yi [WnQ"LN-cV[*ZWbG݅)0̀FQiط_!Ie{{|,n6 9]2 \Շg\**g16ϴĜ*?ˏgr]B.f{N#$\D_-ϳw% w7~,)IѾ+='\%J8US,5qxˌ)+%5h Wj =Tխ !m6/&mH(m0m8ZV6n0]$,ix@+ ͘ }va+-HiyX9ZKцI6KA lӤn IG jCKT-3u9˜j[0Q؂qҢ Ԯl-*ViUALW ݣpѰƓ,\ҭ~j{P'%R/Vj5&{M9,eu6vX<Ɓ$ni:݈B 7q8`+p,nMQ Tƶ@`Ѣ+ۓU2cZq3Ke _l0k Zy(FggE%' MIh&<DzG0:Heyo'W?{`SzԲ)͜wbl||i{RBڃo_"Gucy1T;h;8hV(b`+vejVW[0kL*$fs=W2zeXl{vo#& ؉u+ϼ{VJm%l wop~v޲ ?fQH/ Y:5c7 ЉNo-~=F3v3즓Q^o]lzs1F:IS&sX*kb&֚08Jb 6ZBpz^ym[HJN`Zډ*ޖH+lDj+MolZR;3\AEy}A)VۅIʓ'xV WTGIn&i?:d!W4B QTƳ_($NP j@&6% C9(Tպ[dRG fEKXxP#/7hF+?w1ܣ{T8; uV+J Yido~zwN BXpޟ]oXYTʃ *Q5H"-5|64fkDGb6TMY O=%N'S*O7>Õog؝fΩ6Z >sVcyp۴qSDWk-X`M&, E8LB0LXO )'P]dF3WYC/ PJ5DQX$T`rxGJ8" iTQ) )mjDEq"ZHЪ\ _d gSO9sHHL.PTFPx>t jw$LHokYqFZ8-뚋”5 P\}f=fU5fwYW kYO 1Ajdש & )ԠCX "81twu"p _TP.5YФ0&`t EEF؉1YWv oI+A!rIXh4I4C0;kʆXuv6uQL[ [,SHZ`BU&Wo Rb?qNYvm=A%}ۏ]u/&%\yWT K}A?oۏOG٭կ&pE;:pcUذ,P @UQre."SNNu^QV^>v={ƽ^5Lxy{JX0c[(t2/ʿﻏ$'F=0o倩V{׍s 3!LUB^D^CC~ #<2F*vll74C"oe!)ViwP4EJ@'n{9mY@j|p /"燄AQza-Ԙ#qvsqv 8|ZEELJe:j9onG%?+W7o1\bY.F =B{LiJ aqj͟[Q%W(UҵQu*B*a8KFƢvlUׂSoo<]Jn¼a'!2`L%A:(&SbJ15SR !$.*9PQszPA@VⲺіO^BT/esK!pDY AJcxq)AX&HtYjPA7X Kj~b}X[gd)k@TKDc@I{߶%BP ЗIpIPci,(9u7$JlQ-)u,j;3;3;3;;[%1$]q?WBX#A2U SM"EdIŔX0WMS58T%qdLKBJ&N=RT[}NbEꚁkGq%,,[~%c_t&.NPs P lD`6tpzy`dԆl0nz |讖3{dOYЮPrm6ȃz+m#3& ų܀,Nh(0SsfRJMcՈi#)su09u#%=f2{ +`0|[%rhCbrc44#Z4qZGΐT:o7=FSa$|ݚb_ЉyQf| )gRHI`5Y)y0!Aѳ쌸A q^ܠ9'n0 7X΄Ohr~X58kiU-"< h.g,U!Px͈:҉LP(tHSn)3I3Sʘ,5"":5wLF:^f`~5YabV5O'jw;ü+hR7}gSe_- \v,zJ.1~qcS6Jp0Lp8@+T7vo=RaDI>Ml0EVUI1x2'mryy}nC6U.uɜI\Ł9šȾܲd. QC@[ZKć迒5 !<7Hq8`PuMb˔1]Dm8ͯܨPtF˛J cyc7\ye}2Ai{ T4*_ H񂑔Β3DPs]\LOaևyί[)7>,:ҭٶ9aD0Ept3]SC Z3k`}@A 5 J^'PΗ_aܣ@S-Go>L*MWU!O9ˬ']cԜv l Zv @+Û q سzb=;OZ WJiLZR)Vy本Zp YYF cSw OԲ.UK(R/KU<[h3+ӡ۱eLn\x+C`>[UqìM/^NneսQ+-j[7'Pѧu0Kra;/dx&\׏p `9K ƹbA&xkNG'}'i%L]&|φ0Roh6ѷ$*9"M|X> ӛS̄ CX-?\Xb^jOL/41rJ1[wi$ǝ&u6YGf 2#j) (L&m2^壟FRAQvRCxrLۮ } *H}x*hJ,3DpSNsspae332ZD@Jg DJǻ|i=AȦVw<GI3tG~.«Lz`) hûl48R B.~~>hb|J~B{:_]{Y/9P$@)^|0>Ywa7WC S?oATyfTtOA ŧ;~͌IZ߽z9`Bwl?nb.jGA2/]J}kˏҒVD~RxVUܒl\_e,cT X&#?Zڽ V9J[֪.t-kTn>ik^&L'^aVFyfDl\J8Ì )/a>;n H^vPX$:6 f_ǵ:U> ˓㡑Y" /+ut]rЏIs@ D"d<ZM?S˒X<-ix3GbιUMCόu(?q_㢾O7xhWA ;kymŴ,ki2VǯܧU*~WT<ˋԯE0oWk|w݂ 2,\ٯy5|$"hl2**~z*( gm`vof%ߣ"Y r\_`S-ia 'ihab 26+ȓq w}\]Q0 N层e*#!<EdUh4(K&nũtzEpA dK6V vwk!x{Be愑 Y( c[X=$bLUdѸ3_-^ǍnBL-ukQP(xt:3[Տ:7Ƹ9CzRnY'Q7i~Z$|gӃ/_c;eLY+~B?݌<"p ~^+ ޗ' +I|kJN-p|TSU ^GhhۘǏ_2[0r0&0v& ɐ`8)v_eOVCw B$t|F|-|0|E9^-jۚ}f׻k5'FQ1}?FSes~|N> Zϋ/S,-.- $\'t9 &O&ZpnRՙdp]GYt*{)WnOM/OwhEojKScuF4ӂ_IlmIy$3*2:u,M8e[ 5טWA9&LVkn6 h6nUҷL0%xxOgR\x@H~;aS_?GW>`Z8ܿ]thJN{zV ޳JGZk\J^>؆OXjYA\&a35>E'k/W{+6(O'9lR n@#7sMz٨WxgUw6CWuj9U[$ɺ;|8VXv`~COcgx;n&,Q4,*A,,PI4Jg/;0vwR^˼;U)]?o==7`R x4]gmlg(-L5y.MC-5IaA*P1gQ(arV-oRKhu-C`r"*IS]FRv0Rj^ř2ZLpiY %컥R3X3 M6lw3}|ds"rOm`Rt|lI'_v/4~3'W9e28hU:lt."DŽ}/-T@ ~W6x#ec]G<#o\ZudGSW@Ms_ +netžyзtg^NdyIeY,1 &;jPef>츀gg'}v>S<ٓT{nNO_}5_[bT])1eV*x^dYtxUV h齳okpN%Odtfpt<BF^O/[ݿѿO5wQ"|%aGl 70`)) BӫM)֛O]Tm~7d;{v1~뗐GkYd_Ov?wbRxMVo ؛nn:Kߔ2Fjn>PV3{} 7y>Mׂ_:@ր2JBLEC0K/({xRI;>Oz@_N]2g g4Z4ϥ@<v|*=FUYvBi(&PΐT"x*TlcRK;>]lv9<ʇjۿj[6R@@E\ 9&GZs=hiǧB+6] [zT ]<=P52p .N.1Kt|*Ϋv(y]@>kƲQ\ _fSc3_<߃v|*-f=tXɕ0M 1 ݐ@i}I'BZrq4- y0.gP(Tb2d 9Frhv#TjË>K,Tp_һ懥_zzֿViw)@rѯ5)x }¯RK;>\9).OcWq10ςPXi 21W) ~_oih?Ϣ>>~P =-W]*n;@SwR@Iq|ۍyv<Ӥ]Hba%]<> D#,yMSh  3[gA{2cp]W`R;\A+yL&^󧟣5E8( O |R*ppA t|"r-ά:Qq֕U%eIq@9Tȭ;7)z\VZ>99s"r!=Bx@HǧB]SJk,/9sd9#Iκ{ҎOW 7LW1i$] Fx;!'KɒO|F\)ç\ v˹|+5Jj}L/__Ha-GXc֯#rʵAk)d0"n99=W)4zӎO\/d~~܎0A-foO!8.}Åt|*v9ŲtnD 2Z.*-mI0 })' OZwy(%a{)a;b,1iZ4b>k{sL)9pabzbL|h١7v~Dj,+WQgfB$Ohv}X2'K6:c޷vgI'ireAݭFFCfcTht|*ҴTURA**0=3Yx:%tqh?7C:>]VHvffsYdrl`ߜ  ˕C?ӈ:JqG˹F'XS=]N9*{ms3v~BdI"bЍݤl ,_ }O(pIU7"2 ~e-D"^B4%8"Zɣ 4l9gۇF[2~jY9ZBMJNǢ n2tCѺ /3;Te?tf Wr.gG:VexÝF˯pu-8#_K ΖEs B&.<&koh;ҼRؙy5 q1c\ Q~)vن.f ۝}nPWϒo8pυw f9.)gpu^nfmc1ow璫įomHU/|ONSDs{syJeS7; 5]|55&%t]38ƣN=:ʹaVK!JE ֏ROԌN엝^J0օRIKHb080@?t|*R6&\i Zļ0~a1t,YM*(h}3!/׶%󈹛 s=޷s~ޡΌ6^-[٣ן9 `=~rҐ~;Xʘ0ɡ`͠:NҪdfu?HkH'mfj⎵D3㸄O(L dלrߋpjƨtr#1~Gep~e7-f+Np{]X}ᦄi';0u{( lpl1<Eo׫~6}e9pU[#NI+΍~hTPgT)lbW Nop^E 985xXVN j}t|*3O9M[My:W9|1< 3~t|*3j?>hi!gѡrӅ' 96^]m`d0![TѤJA`~XhԎȴr͖v87RiG˱&- WW7q  =cJ1lRԥZrO7 vR7UZi xOˆ&jj~=z3&ޣ")c/R(OM 4z؃=Ǡ唬e9-M*5-XN?K^ꖦgV&akF`I' 8[%kU8\Q=e5b ez gdKMq^!Kޜd3v:~L_ks.~`hSˎΈM ׹u>i៱ iBw Iъ `QRM<ގHjShxB77Sf ϔox\'zԤc I>P-[Q7lr! #IL4):̣ |ĤF)2,[4Sy}@k> ºǣ N~&f؞#Ȯ.;.OnjT;}y7.zU9V8"{!<ٯ!j~[6?+Ǭo\e}#Rhڛ H9)=g7_|5z{nkxQxI"fLOfS"bj̬ufy{l?Zɕ뮎 1vLkE!"S7w4|q߁jõR/80V iLHzP6IVxlhN0iI꡽)̸ﳅ?$oE̸dÈٳL,7,I=_XYI4NҸ:WuUԡ=͆g I˗lFL0:.ii)kf'%H;-hЊ٨|0C;(uTt#^Y A'(ZյuƔ*+,P 8@Nчt|*7Á4K'rKztzx4$2͈rxw\ ֣w Mkz**V(S琎OuM~{G^0'AzP GJRO#휋R c)H"S! E6V?/cp3f,ʛRexh\1t?@:>f9w؇EVuFHUK<.^RrJ×7Hl%ZIN\ԙ^ 7ؓH\'3kG7b`+Rҷct7x:T8(J9> e?I3?l?9)Pz3u{{KHK,WzT^uXڮ!7k`GD#Va)[2 Y`?TrtcMnRNeG=c!v^m kF.mM\v6=gHRFV v ЮڄVi7۞!OO|X{IPg6:Y*{B5$q7hpwrF|&9쇼 ;ɇJ={$[Rے]vwSИK2X,KMM3cUJx4g]3STJgUjpI V{1^`zƅ7P%8oLP팍G,gUs0PgmV5a1S:ng^P Wp ϛ_3[czǤ\L *_7O0[7 ]>$_:$1ͨN&2}׏gΣRVxL8eE>8$I iG6]$ ;cI2:UQ;G(j]ϝϋyqw w&4!:*{hyI™m*b tnaV s8ͧri77 ÿ?e( }ջaOoSm^ bq}%ObP&YD8-)jc32!౏Tz[B^|quw @Ηl_eoVN6X-YM3/x*2"Ȁ1ɒTDyMNqNDߤ!6~ޭGAܲz2@[#|oRR-j &nQuة E{_ϳ?Եn_mwڽ,Z7Wyϕ~2'Zqy- Gd{)#{c/ԉHULJ.7QvLޟ2NMx}_Mo oS]/F;Ԗ4yMĪsSҊf7QB0e7/ i*lbAcXfdg"\Yn5|_;(D|#|c:}o7vwg5GäyAꧮ!#w|_!1Tl[YnV?sW!*:%*W7.-YQCPc:|=yz8ۢ~P>õE<$A</oڂWQ Q1Z9Wb =ш[D1DQXDk1FATE"╉`G̖O 7Kk81xѶϨaoCgUã@) U/sN'4mOqPуA E<֍lWl8%僐|0C # L8`gNH*IX$6rFYa3˳+f%RokD{8h3+9O쌶Y1!>+qp:漼8VS,ֲclcR-#nyH;#G{ ".%PXȔ t$#8ZJEͨJ3Z[1QXEZ1"e,sUPtW{K3`ZC|M?d"7lvE_R׫ro ٝFI8Pp[46{Sf^*i0 -Xp^ǔ zP6Bo5.zc7ZQ6N!C^eŶkmWCZdz4gg[_de1HF:4bG]q`Tmh2&T\:8} F9uĝvo]FWص;թ!Zj~k"=<]]2:R?!Tj[t-fPăVfyxLʐ*!7Z(H`*3ӘMn“ ] b!><ۅѷ׍[$j$ޒpsByZI΋IČa'X$R&ܮˌհܧً!mOQcݗ'SMlLGSdP5knk0z4T9)T[C9^b_1.8u:Q` DұӵPҊzWw!FVbi(m~QT[<ǂh9}B״FB1Z S M3K} F> 󦼂U,bor}6j?.;#8P3([_ui:KHb`Єe$vi$Ux ̸m/x5}u5V?5HZ1֥:`~HPEjcK»(J&CT{&xc x}FG1F}[G[Gq93 p1j?M6/O\p^58{'J@ iQ?&}ҟSסُy|d߸"oiMb_1,R)8ܕt#ovx 36q(*!*^XxmxL_vk,iBGEUUn⩶|GQF޵Z܄k@6?7],*)~ G+WBTa}+ȞbiSS7RôSB{G7.qmS^^8%` 2X*6gS^59GJH$C|2&\epȒ$*dIC<̈́XP&3Z:㲒hi-`HƑ {Ug l *FTˌcوʉcZ/9#iF2Q*ׇ7-FM[~bӕ+1ư]l>fh(=H15՛M ''o;wdHh`w˯c_+ES*yy)`7JPwZʖeGDj# Χ\4J#fvOJR:T ql mD2 IBȬ2%qے  f3 *Qyz,Yq,u10X )`6%"$eQmF#Ye981'Q5. ucsc4Yi_jkրs)M8 &Sۍ0ւP5 1CrY60YF-;Xz=#ޡע!FǢUbГ=DԴKaRȴ4aypk_j5{*77 r}K^Mv׭˲T˻t]|-r\8{W|YL%?4z^釶߶.?My3EIiXDRG(+iښytLS1rBaqf-skfB!at6=.jRoƧd_ë*e,rXه^k+K P,ܗ>o`)4wUS{z7zh$Md v4#Gl% gz @W bV"Tevh%LR j`$ NdDJ3=U&m<w_{gLFY~j~_V[O.Wv>|;Ju'ugP/}EErFR^)}84!?"N WǠ1'ĨPQZ>p2Q؂iZ2]}~hK+ٲARz~ s/8Z+ScB4!#JU&)gE7@ ( T@;ղiv"k Q+0rk80aΐD+X0><3z'A<JMi`Umگuݖ&8\]cB< bquO~B+̚:xdn(L4JMn%qm%ۄdZ[Ip ̟ G !b4c "㊂,pЍB<\:ɘtc=*[ꘒDШ  .„rO eB]9Ai B&Y!DY1G ѝ;s@ x>"H'$7k1(KWm8kis?^-?ߟb/Ϟ;X 5Ϥ q^u9u0[.ܭ|jXfE3iߝKeiލ|x5c|_KԮ_8~ȃx@67\kJH$Nhܸ8ҎT*g6I{*a~M#NEX)otQ#~t5z6 ,_bы?kf upU˴d4{r_y)fϟI<<'K~Îڅ] zTPh޿JɖAn׺TSla7;8Gku=lZ#\6xIWb~n~x;)l77?_꣢qP*;^fwaeG`p=sUʗ.ߔRItK}t|aT-f/$V8)?. ph?fOP, ȽQ^ ?n:*h޿<s.OGIܸK*ɘEܚ%RX(-)fߗ9XkZʰ2a%&^YPI [-{/D%9!BԔX<9))@B-M 9*%&?xW *O{ πw !'=y4Ň`_/zdMoon#JQQB|+Bʏg5000/V-cY`nE8dht.>{9|Gv>ZMG,IUMI>O^{ bz0[9ERe<@Xx$.K-v'myА6_`l=.j4'^FKPT\6W߰~t-rʆKZG5=|9yvˑ7m2}0z66(GQ5lwd3E 1D,0˗{@ћ&>0I1#<0r>/eP}d+i6=,[CEO$wL|!Hs#P@V&KfCU~ xЈd\*f98|nƈ,cJX~4H?L5&)%`F`A!5njr7LasT($2HAsb ]TPdEsӖB `3cMZ<QKEz.~tޟE:H`tݫr4G"/@MU7bubX}xT]S0Ns2b ,*Tl= ȗ3"êabހwSF#I ag}縄d]BIƈ Qg[Zt!9sHNIs27*"HdP@,po ΅y.Г]~pbj)Ƥ:glq>cNDo9:yWuzI_|, qu (&Qs2.d &>"-N={Gߺ!1[yjGE;ݵ,`!,*jU 7ex :@ھ!NTcK&,˄*}z/`,c 0A2_Q?`:Ӛt:icji颰x5KsڪWSu,bӥ<07ױ$[uJMr9}G{,OP|;OPpxR/cD~ggY4jZ Y4;V-[W1q2y"mAm .V 53T%`2{sd Qf9;_IHA4F$CFX~}DaF#6 2=yYx:1}&p< Y8^s(ɝ)E3Fa'z"۽Sپ$T~$'T&X0CDy*e {;fT1񼪂pyzaGmrxjo}<#&mk˭2K#QBDh /[X>?17m۵E=zÜmc@`ם"cO[Ye_.F%b>u*ՆP?'ʍ'<Je_e֨_Շ]q-,ApPHRl6$T$4QNSn+< B8*`D <~q-j;yzåۼדipy* z}w3UGĕȘM#3Z 5&]y;$jM.x| e2`D8p%y8,EI):Ƈs ok;Jd WQXw sf-Gm/5SvpdCGѾA.ƲgS($oxyQbc=BgKQ8~Rd 1P0UTCo e !4QN)bjKWWaM6/ 2|Kn b iHfhKi#>Taҧ:8(>ؖxv0{#83B#j A}l% w9>M=c|dXm=w_M_>;nߝ.x Ih*ڧ0@0UΈ ͺ9KkoލCcl7 ^u:Og(} )ٞmK,2%,ODTkd '( E8? 17,S:әxlP4/P}k5X8q`/H3'JK|=H$iEe&8R%@\ T3T"biPb~VY#leSww8ZGx#s],| $}nOZ׽6(FW-`׬^bjG 9,g)7+Jڢ)s CB'vjZ=lBtH|u`x z* ;^ ɛyvμ;fӒӋDnpTQhp+ѧд )erc3y7ͳrv2ڷ^Mמ_{ޟykagn ϚGN6NuuNUJi^J1 ĉ T\aԶstmr%]~n~x;)[?>*:Кz'Kv}P+f AH3TIZ1i~?~sxQ A,5ӹ#̰d2xk,IZ1a/P*5ۓc(ppNT5ÿ;?o^w-ex;f[_zz<3Ȟ/0͂Fs|6vE? /c]jˈ`DMÉ6sM%r> ȾI_N hn|7>esBFz|O}Evzxӎ81o-jG;mեv _;%7Th ֠3( KDr܏05|Qh˩ } )B JB2^ơ!1Y3 n&7ҡ1L]42L1&=`]${B3h IDңǧ<\ZE81Ǘ E/T1HeJcxP h2S2){(JU%inM+?$V@KJu3mރ6 }u~Qn};jWWlkvsѺ9cksƺ2m8`ݭ#x?noCMA}uhNC_wg+&Y~u*j ջEe]B9"7,&ϒTyKiG >]jnS(5`]5AڏRW:4c狠E3X$V/\kōK2U@P⏉T8BrCK;d]WK[n"h׹:IBCJ XF1U*-KP!X~02'^6<ڡ'2O`20kO,jJs,]9*ΐR0qOL} Л ow!zj NFCNt]L Tܟ~{ 񨘯{ÝtŸjn2 /KOO:[߾>$| ?,h}óW._9ks̤&ōQi엓HL$CBU/SYsI&PYιr_Mo}I{N(7e}%:Qo䄣2\7GtRNοd*ncĀ4ύWw@6S&rH0'SW5r(TY;o;=ASށ]ȡ4 qԻfrX3F*L $zMlQ"(ixGBΣ3?6R/@ F^Pp}Ѯ1R7DŽVKxܚxhXt0\17u| A9r.vS{URT K\ }&0ZNf83{=Ǎ#W0_;C3\ D(LIҋ=̬";HdINc% Fg9Nj :M&EG!*>&,˄_CC6GS+HXdB! BU0xiP5ӎH^N}[wөnރ +Y:|CH)p8Vpx?2cPܿApS@w#9J#9=sBz[#9}*Yb23 \&sD̝HJ9n{GsODPLщ9n_&1IAyKbEkqf6,L%H X <aIDuT |F\%du@](QF[z#GɏX:P9Lw*q!_Fj;ywG -HMS̜?D sRb2*8jzccؑ=8,U n]ujin\]S+#|%Skue2ӢG a'9,hHƿnz[T+o:{S5=6VV =-t<|z72-%?@i;޵E06ߏ]6A6>3HZ5c;bKAV IPPjXr(B\NexQR'}7'eC28|"lY=4*St`%kZӼiT0TvhT d]8b, (,G e$4=4*S ܚ,K0٦ݬeN햃8Hex xfK >i2*p}9KcШ +R@S4k ] z>@81b)1q\ղF(ʘQ[M`;nnjRѺ2\0 YsNFAerL09:+Dar.6s߬M.CiSi(p9lEӾ5[eF[E&40$'xi(oe:ۻ3vaC*RXIXZx15FۯK p]UXI!Y/_Ud&9QQ8]Fd6ZnIŋE \1dn(h [}\̽hЦʈ3[\NGrܑ-pN Y=4*din{G:?&<┞!R(iqhlD2Vb``n2ﮖF2 ^J+^ \˙݉cFZ HbZ/r7ԑ'c_:O탬*dIBV(a܄Dz8Tp,>QДp%4 8 w4|!`9YZPy D vľOYӕ9J,( QƞK'ڬs|W-Fep;> ֶÒ7JR1%#5^ v}{9FF$5Rww(t)5 KhG?^ߵ2V4Y.Vͳ+BD!X_7_R*=XG^,ᐬWsC:G܌#8*0橅Yaгnjmk$ /h`:T.Ŧv J!I% ;)=rmbugF!KzK\ot6ueÃqG]iO~{Jaf%uڬ< ~O׷%Rr=4*SV>](CC28,qYW>Ҙg6~2KsШ (/^gwԬppWzH!O15S؆Su.6A)rШ *5VqЖb,nc/R$Fdf4RrEhV(blI2)NKSRê1skuYօY*V@SĎ,6QZ~繓35b[5{hTb;8;8oRlYhШ 8kXVh"U)bYu< T[]~GQo? #e{h/l*gGlUdj.C85BRdmt12̷ ?oy^' 5p:N/|[iIiW94Ҽ"|=y7?m. ^}5ٶSf7O?j[n?ln$v1,inWW04 nbNg4u^lV.N$H"h9\aI}uWk 1 ~>L}GYD0JFoq*]T0cݏ̻?H3:4>S`A} S}7N5,n*LHo0-~rwK;ʝYR0I_O& wrfE8۹zG}V>CZ[Fov+v=$c X ]:C봇yskVoV͹~.t?x<+%G*϶~۬ؽݭ)w3T`@%m~Q5 !g+A柛٤A|pUlOv#Mo> v~|`~أS+M\LCG]/޼DF_hhW-Ŧ}:5MZ<~"9h_@Ӓ9AXT+4"4='盤3ջ\¼جg'۲P|oJ}Rw7m:1zQ/0̷`R$Lv2cھwf&Tg!,VOd;{*]FO-=(]Ji<;,g"%lKUڣQwۿwjҫꫝJt|߁Zص:Ϟ+ϻWn7Gx-{X=ك[^~GݷvdenVv۸_nul$eN7u;7&[z8Ӿ:[QVW8a;{Ha{ qcEȨ@{ȡ=d3='qeh^W@`RGPJmÌ6>6imw1"#(?_#fL'Eէzw_镧jڤganF7+,`p+3kְ6U)qw˦=X[Tʙq!P'%4Ame'LGNϏTl];\޻'L-L\P/iP{g|:zy`g<q\~k 5_?g<Czdd2svo[ a**g(%\ŀ>!LPdȧ"F RSnۿSK$ݣ|}Oz{3k[L2>lPh:}w$8 YPJ7j)RQ)\İer0]GWJMZ8J6 ZO$Rxc%¸c_a6 9F/}k`"{&CDisE2[pCL2qu0f[N([ 8(,X=` ,QLƌL`^ˈiDk45[!-h՚B[7sPx)FΝZc$'2Ԃ&ŃdL@(T!X2xK]bi :fNhZGb4:c !<7cd8aN4!c/!l=0vbGR\#P&;1[a@ X9CBP*U VǨdt:Jä1 L?A*J!)RyB:15H}Xѐ8L0 . %5^˴R'R:&MB.പ(ƙ% 8=64) B.jl A)O—PG֪egEEG!/0dy4*Tz0k i¥+-,(,F8B &x>-ok0aShJӾ*΄f )KvI { q =-Jp䁔TAE"Ou~Cq4((Da6xZ*V%S!,E7ڰ"Z8T$EAҸZ!pFKF dLʤ^SM$6`SLRqnUSMX{\596@"KN\Qs(J[2JC) ERΎil8!oVSY2Y2V(+,S;d68IhJ[8|Rq?c,n4K丈^<`" P }+)f 14yjƐiHY03uH*nF)K<;[c tX,50L|DbP*c*ʦ[C':ʩll@Ȇ*9nLWn_iJ%6KhS1Q3Dt0GQF4] \QGV1(h+:~ʥi"E4VHwW&[`R«qߣ.K&"yYEBڦNBI BD A %ŘTY)4jSo eXxኵ&kE\uXgLӄa >=nm(F%̷@ M&jZV"!>X\lrM,Ѹy4/. D,/x7VTlh ^f,:;*-M DUĝfE <-&sJ }YI N;>lRtMv9TX] XצXkC^ 159]rmԎl\P1A߅^* FYuE_Hn&)b,C>j6y:Ze M. @p Lc hwX@zT̨ D#+FσthZi298ksj#:& gh `?AjءwG{cvzb,:FaE^Y`h#܁?Ufo'R% fcI; 4] #, f5*)8R:UA*U9u[zˬvF +렱^@̤I@ Iު0!K֭hL69Hc<[`tZ}OȮ_mmU53  DPC\tBÑ$0z`2o6:7  6HS݁Qfq׺`XCm58EMy(Y5X`4vƸi᧡=7fܵ9nJ4ś !u>fTsC5`<"d5h+G JTF˄Jԃ Z`ѐ*<@%IO p'r5l]߸U5úec4BĿz9_a+@1\ƒr>@b9v%I^g<2aNV(hmQBaT(>5#>j CW=kU4~ aQ=+5"5R&DndjHkԪT gF (mͤXLGjJҸh vh u~9(Wst Be_j;J*>@贑12ČG 0Ҳ=f@rem]0"#]2Q{OQtQ,6 8I*t+/A$\tksڌ!$rU.u/cdiVMjS‡БDSZ,9vQ-0 UZhui]~VW$<Xd3-#Vi]ō7ʋ6RgdrYvxpjt@.\Uzfz(-sw*еN il m>eqQ}7ŀ},zEؤ SPcȣ6Њ[ ͢[]+B&!?i>W~̥@Ową9|unٖ߹#mͽ<(} 'ib w$l&g.ԶnRNgPቃ8}I29G&Ǵڃ1mJkH&[ɱNGq~rڞdc,4UHLl.y1kY&ގILf2F^E~.[?|~s^wFtC7*WvgqډNv8i'N;qډNv8i'N;qډNv8i'N;qډNv8i'N;qډNv8i'N;qډNv8/ 'gEN˼MBө~]x'b^0D dM-&` rԓ#I^mowI8_(Ed~?Yn>̱}; orp$n<Wv_zʻ8*F| > *J~\<eyh:7G8Ng=.?ud~3*d'6Wo;91|ݕ=Nμ42Y<ot(,tM,xa#JѲ8o~,JXlXd!{ۓ g=4w_Kh_֐lNzYA{')Uy< D. *yEx9ꠒ7s ,!׬$_W Qr IFX1$s*̷MZeւmNd,)*y]/Ӽz*(y~Vj1񷖓 i i i i i i i i i i i i i i i i i i i i i i i TAabo . pdkL1Yu& |pƸCc;o;ځӈogMw4kiMnw 8q85>S$%8L{IY-Q+)teۄi#D&40Mi"LaD&40Mi"LaD&40Mi"LaD&40Mi"LaD&40Mi"LaD&40Mi"LݝZܿ݊&~k鵽1)inIOځaxEV Ws<^/WdxEW:/ONsE8&\&6# ԲvL c6xG޲ZX͝jqil?jApv۶Yz yh\to;X6MV&۳ l2y'{Voqf<r3WC Z^:XhVG` ]covgeK~۲@&=8UZ;~C#&u>{7vuw'۾:^{R~K(wJ;7V%哹ۺzHݼ~|Kiv~=e[9KO=8ke=:srԀ!BJ"S?]:YsI\ 5`] r[_`ӾlMp>&_ď!=(`ulxItOŵ#?h+߮5s8qa8_]j>xw'?ُ/ՏA/0d9/ nodtW^w75A6ن5ο-85־@ZYMo6if]./;䯓UCL(hyAVl+"s 6"_np!<@U+ ):`*)WMA`ʅ%#pɠ"Bꦈ\!awX((ƾCZtHHHgt!(V ߵ)ޮ <? K\1꾜K ޥv㍯vKE,\ëc} "Z%^y(X)$'!f3Q` 2<3Yϵ5Q$. AT ,< D "!#a!`ӂc@鬊)6K ֞A"RGLhKb 0B;(vJr. u֦skt؎E&NAYш^Xj`;ٽUxu 󢹁cG|~0r *tJSPz 0]R QhPfA ܕ1:.IpPq_Fa5x ;Bġ׽ȡטAhf,M3.<~va7[LͲX۟W壿Pģ@6M5A0#m/,hpƂߖkmoIymq'`eX{gah\V'H=#C SZN *)Q4;`&޻iٽ"˓>e" P@5bxfmt9;f⑒. U b*XP FJ Dgjb=b̠EŖH!1C{ľX{tN?i_1pNH]L ond#Q҄%Bf_u& Gn1߬VQTM dTզ؛~M *Ҏg}*Ͽ]N!d --7T9c"BOޒrʿ)g8r8==0JexbvȮ) a|G :vwd4#1N<.;Lk 5*8錅E5?A3S;{!L{8֒ cGt {︖D<{q;m|o~T7rۀ>PAT_D+8:X61yχ0aGCu GFcto< λoa<D஗fX>|_})Q zK3:s!!җ! MI6dhb4`|pem V([ WlEU,# b S~Ơ%QovQdӚ{ߤ gna,kb,k$|؁ K>q赀_ Ձ+.[Ϗ{-pseC~ S _]>!ҟq/WxƝS1B ( "8: ,NdrPv{5椨װUëG>eWhw_Dс휞r ^p&n7vd,! ?M_vjB!TOGS _eE~mu?ѷAOIccS)XE9c0-!JBĺCJ[bE t zVůV"{F\zP`uVpVHe}[--A5~$tߕg;7.M>;Q3χ77lB J֫ŧo l!)aY-?+ .4F` rl ݆- ڋ1[?)d cf%y`. HBi9LjrB1bT!A0s4YW*sY63>u*>acy84+͜z+9~N{O.o4 ڷ=d>KҮ ߵ+E#a߼T isˢZl습b\@D>"sz*y 0˩"F .Ҩ>݆+DuR|IF'VQ+mWFnu~.fO= ̰ o^_}vX&8xD9 y|]1_;yxKĀ̀_ 9>p1{d1c33g/J+iTnN<с ٞoF+"妪ez?qG^HQw,_tEh Sba5z2*GNgջ5 4dA^4zƿS@"XV\rܹZ|7[HlmFJ Z5eL1DAFPn*gDqgdXAHA ڭlYzjVW+xGnV0TnT5<ؓ|T5( qÈfr$-ZBzD#y<0\Xv¼]4SƕтB԰BLYd0 9$ {mPpAi&X+BB,YDzݐI`QCl9A>*cc' ,"+SRXG2kZִiCjڨ5lHخr!ϑm;@Z) ΆTvqVAtO`I. ZBp(\RC>MoFFLBXr?٦q \ iN I$($+{=$vPp NVv32V12rg̬gС^{j|fUny`K ow'˧dƷ f w6Aެ9S&PiͥoIwHr>gY>o&wĸ1b6ǼɇɳѸ ) )% ߃ɏ?||+cH$ ήSZk88*Mr˶yS?m6qv${Hg-앍{ד5Y5,{/3bɱ.3b}X&sd>U[^ NEs >Ë́^HR6W_+6?qɓG^NvP.cd!]]<)FCz(M0^>cv&O' tq@xUeU#9n2UBJhF5XV.@$m҃oS`u0pɩa`sVC-)~Q7w^uC ^rv Uz%/.쵑:wíHn&R$?HkiX.._ ѕyT)鍹)nQx(F>²wd/"6×˭X> 4BKY>Aos1À\ 3z|3V1K%88Zg3z^$*r+7V)^?C7*CDꚪ{1r|pD]jiFLεo ofKp4oXC)sZTʁTl?nZzo`mfdڙb4H}‘[$ʶg2sL _ !edǝ!Xsf -Fl   Wu4hAJ>YU|$^{Me<1ɪ#Κ1t[>'4NeυR gpn'@2 ȵIzy$_}2dwN qHIkc` ޾]^O.G[6'}&F[jqfZ@Wt (yc\RY# ☨H"W_2oW3Y:uֿ`n W50 , y Nyly0IT B)d R_ d»pcj [=,O? (p:[㵛g,^z-I *NR^;0Hx`0w5 B[ט rU4V7!^y}hx}țɰm>\ W0hBOn&*!`XEx`(jSÂ1¼*s{G;0 +\3گzA]m@N)wh2/ѭx[_k f1-{ M&80{ 0@ 9WJ*J!F /\B=.T y\X݁$"4:xW=:D%\Bo% er5]{ٖ{72ьg.ak pƺtHCkW: 6")c +b$Lx= Lil "Y/#3\[SER@`{( Є=0l   6SΊtvE_N gmJMao2gokm)V\YFh'NI΅1tnmΖPv[Ns޿&TCa8z)f8L>x)%s/O/D<]tpF>wF4w Xˆ !MR23v؏Q{e,MX?GvZ5ޗۑGPYSlsOf2>#/[ :ޮ|3-PkQp'Z'=՜= vǮhd 9Wģ9Nl43rm@rC<^ ƙ^RL3]sI_NO>\;=i}μH5*a;w6H9 V%+]7% ܄gǫ3ݍ\i}6C'?\f9_,[~;B@t+2<< 6hVi.}[7&q7ΤN1z5V~kK̻vee/H{P @;aAik0 XvbNrXm'9Zr0Rr0b{?^-^;`6=06B`"RKhg*Mg31YǍݣh8r;3Jg,9RU13J 7)yȫ@p )ʗh4/)PMAmܰ,.AX+@B"Hakr D,Nmb|)kލk7ی´{+צA 樓So T&*:_I keZWk5fcsZL&X i,{T]dn<~+ w[\W2xxf{ch{ffseu֪vO+Yd7=mƂۦwp'wvmΦcqkDPzS咽W]w7xfO$= }Vƛ|[n=Š[ky!<>ٲ= /g ae;~ yͮ>&MNғ.VyY!Js_r*C)}]gv3!Lr|Ay B($u=1AC NV7_u{^"fwW9÷V?#x?_ZSik{ی4*Z#/&X\,0 YĊ/?ߟU+#>•짻nMv6ξ:qGXmfwSTw'b$qmQĬA˞ Dg+ Xcs,p^gjJރv"{`m&y1-u4ּq]fKKݗU_w~,W Y`2{wZ/ه?d)k\ϮNygٷtiV?v B{y@]дY|[aKxkAݍ+m!.ElX=M@H`#RߌBϨĿ̌|]Р؈ƱF\2~f=d7 L3vSB0B{ L"8nZAh0Ln?~YV?Uz>4/rI*_(J<pD@;WXF T9PfnTj8Jzfa:Rη̏"͎dr3ճbzs;G[Qt~/n:/y1=fQT"LyM"^1V-W'-l;ۏwEkZa0:f#tZI}#|Lё):2EG#Okf`7[w7&Z/쓙e &J7@`H.:!جl˵˭ t§}'oasD(L.9CV p6TpS%pUV7!,OAƴ=M"6 Հa2>ugMIvBZ1G/0z6>hRD+\ ToѵZk}tQ4 L! \6UlM |BV$rx!h;©]~7wV4M?-hyth6_2}4e'g7Al)83EҲ"NmP{tBAVZ*f~M|B=2\>ANH 'W4M:1DOMENGϰtz=6;dJ`_g*+C' # i" ݌aM bVĹyQQb?M1::LUi)R.Ɨ3lbo<盋hvtڻ su!Uv!M Lk-KrU ,PQ8Ǥa0E4NJJ8%Rb3 Ja䘊 > Ȫ;Cd'Y#'ױ-UC8=jN vj|e_>w|MC}11Uy]G|}Oon<҆n7m֙ME]؛\u[&$\T<8`]IR1Ǒv`m'J-e;:I諞Vnu1P #I߁\cf|3гkX8t_=,^SNܗȉ+ us|y&Xq \r2E &Aia riC`F Q^SC*/7cE~+:]qAʠAG:z.Ft5Diꅜ+[u]_C'_T 6J![LnLq/)&)g҆#=0gs|9"#B1T) {)G@U~+MRA >HD*dWme K`*7*9^*hWLF\˭Fֻ"|hg0J*mhx0ނ\bhO\783^z~7O__Ҥ-Hqjg{]9o" |!,ȵn s=)_ʈY8L0 ϧ. v^y "kY^(E,j Om@z^"|qϤ$` iFè/Cr`]S;UiT/$X)6\-k׻CkXKߡj<z!~% 'd^=C~Xf_vI9gH(||qK3+’9|#mVu)]ߌs-c`\KVXU’oʼX3v3!K-3\ ,ɘSsd:YN1TEK:ǽݪ]' =q4[_1Inׂpik4|/q5e% \B4"nZP:RjsP"ƅC>yjNDlz|X.~VS'_m?䒬7<շN;EdhK8,_lOOAJ8\!r -%&12~󚻔Rmi=kOpt1!* 9b+II S"!& ŜyWP æ )mΑis{Z 7Y>i?IHV_˩RKVD!O}3#)u.6 E=IRqKK~)2zPF/UazH_ LZyD^y4vaӾ 8I5IbXji*̊8wDW{G@>gG-7Fp<'4Wz>u&^?'?[Hɸa]4KKC6 @NcBkõT Qun[N/vtRj Nٟf3m@UjxË!!>/~u ;-:&6itA&ʼnSȾ h#)x;нt }9xGX>`ǂ8ݏ*ӅG$7*\||K}S mx}&Z\eƭxrF$f"U2/H}6^FZQ\=(w$}7Qge!4yYʓ;%\x6d;eLYKIVg>1M&'Ǵ\% A=ō:}4NCŽe㇈\/O퐼sYX8eu5bcCg9g) G\[# /VnoC>W{H^|-Xˣ/!%Fi[':h 3{*G/rQLQ++rS@_Ye&W3N?Էj*؍KB݈6Ih=t>vjZO<'f[štC?Uf:jk5{3O`c͜B}%˙{FG:X⥣D%r;#7˳❟J긗>Q#PEɵ6)FL@O|0<ٜF4tEDF`hd1BZUR*RJ.tL9hv_TDۓL&DMhif@EY̱(yn"2e0+#ƙJYF Їh yrp^SZE>%|4șj)鐢җQ(8pB]o'\˄lR| !./JwAD*d)u^ɠSԐj B*O^PQ g݌ ᷹'k|OI CDpNNj\1kj.DMX zKѕ1hiIZT܃K57[P]i xsKX [Lhq:kVML QڔA3 (gHRIR(P[aLL5 Am5"dpI5J_ xzK ROMva! $>GBTad҂&4cJD@L{NJDL儀:-ETuM9Ҕ%EB53s^ |J`m-C0jԻp?MDmEԇn^ I2&Sb,Z=@>E(Eu>HTKe#5thulѻbF d*@ AEZfr /VInPR %w8)NBi#ṵ&*5t¨7QM-c])|!iH,qkxʼF6n Z ʏzdΘdB{ ,gIdZ;gt< 0T`8ZX?0*C@9 ( q !Kt8Zm0Zy9 |!f)co(|Z$hQrzafUݳ^ hZ]ާQHۂmSErؖ;nmy-]1Χz풹q3CC7*DŒAhpLrk B?Q\CH_xӸ)'dbZ%2v XP%"9,Ax{Q)Řt+8Md߭3:k'VwM;֛֝[jME_ߺPKm̱,|୍(bK z;7:عEOWmpK~; e:qɪD)^72Px8߾31 :u s]K`X0I,?3AJ $QK+d8A#Z p$|@g:N탤1UyPKK-8ӔvLJbNmAULM*9tHUW)WRDV:+_ 6 !]֥Hl9 Leh,e40Ϭ1&d?\+MEmf}AwP|run']ufY`d)| 7:Py&Pi~[EM um/."?^4^ ZM4^DːKCܤdoM)z`༷ 1iVg&rd.{g+j9 4 V]N,;i{Us[=SV5[,r%Uq {23?!G]=MwOL:'wSG9+|]}1b3/fww'- 7,"{&B-Z1dwYsMs6O;zY9 ZkZhwTCoc_u~wD ڝkا[V0=Nfvoa<~Ugy03e kWa4zy]/Q`tY'7\0 Y#idSkplZp$x=b<;`.RV |' s}W1Ь0NyJp+*;xrֶ;M F]Qy X>_X:8#?GZ6zTGSQ~Q-˜HHbI.'^zK2^Vj 𒱮@;t .)Қ I*,K1L猨'hRhcWA/9ԒU& ~׷燲ɲe>Y)5lk-ׯm`;w&')0zӄ|6Cy46S"M"Lr @w}Zao]kƽ6]9`m.s֙*ѐKx<~շR Po ;JVl5Sڣ5ë6O ;@f :.@a)Z+Ը1Hc1멏c D{OjrNQgKD6ER-[hx7;Ku|`Nk}-I;{.צ]kny4jA P95Uؓϩ$o 7IKɖJo%jS6ʙ%l ⋨ 5|\-{{us#i|}ٗ&VOovKVo-ș1Z  AYς$ s^cPZ`ˆ`mR1v7"[^|<)U:oK 0ЖJ&\P; jGhZ{.!Q#/Snɉdr*/C;/~ nS3%Py%  v `#_ \XN N`Dm*`iL~.$I=djя0 FWq4ҞP-bVKi ^#!0f DlъOq!9qDidp!"m-V+lJe@jUtCe&0'owS}V5:{O^KсG?0">c.O?~Bۨ;AД#.JxA8Jf_Y)&g[W)8VAvÃ(_yd^nMgMm+ V-xYm,x7 52{S=**e{Vu~Y~54ռ͚q?o^g3 ڎ̺zq6f_11uAV >}BI'WRʣ<)zʣ<2%Hb'.QOySG=QOyԓM S% P <3.Are"<|gb"q#MqLa b77e;crnʆNmQ.:F!jP!Y-sB!ƭ6ѵ"y`pڈ2FȍY[=cWLxŃ`U0㘅2;#.0^YDar;ip؈V)xoAPSkjqMS0lrCH*(Ng~le >͵70E+ 7LYGgi86T{G((!AzjlU[)^ZQ^䬦T7&gNQ bR-(aW+㒃A a 3 %mx;EG*%fOM,X*TD/KERTD/KE)CTD/EF"z^*"z^*襂 B*+KERMERDcRDi"z^*N[:ٗS!jh "Ǟ)BYjDIw3>gSN)4 $C֘<`sJ +hnĚyf).yĠW{,DR6}Rܧ]"Jx ۿzB wap'ᴡ:vs%XK AR8<<TpQ ZP!Xv#%CQ^3.Q v?Lpg! _ыu x>yxut>T>y cў~ 7^<c)\NSusτuT&2ePOfPw*ThlX` f a!wii=񷤒c4V˽N;08,()Jp-ft]Kͻn/gJІ}"tcm: :cj{XpM6nǬ^ͲChC&JsnW-'w^lT/ydΛFtfȳ|Mù<5hXCfM"*zIőg<}·! ,-d6 b?yܣO}ԂNyP:FzUH,g>]mK|;cAԶw77*iPc#޻gɆW&|} |".jkoU붯WI_iL9ʫtM6݇NYFq W¦r |Vw f%Itst"!(|sw0~/e&vz_r+iJ$| ,D (3f"e(t ]]Q<1QLC&`ZқJ6NzHm[m6gْmE=^'s'N]P|ŤT!k練6Ê +=c|}Ml<5-u=T5^Kh 6ǝAy^3ФKⱫlthiWU6}m}M-/V Ibr<-jNG-9NZ~t< O; =cH9~qV.ۿ|?39. ˨4Yx14<GH:6Z sכ]`]<Ċˢfɺd[fz |]6ΕU4w6 dp!)%hİyˈ[CsSr۽s  ?Y,B0lL(AHD5 RQf <',R|}X" *#X7A 9GȳVJyVxs .'æ: ӄAᗞMTiBJc V ]Id`{hUR" 2r 5K Н0]$ғ.\I0i&/:X k+xyR*bQ T/1"ݨefė8twʽCJpL7L/ǢNWzzbŽGJtퟪʖݏ^LuCd׀rVku.X&0a\@Mv_@؆ި)aSⲙZc)1|ٜϩ"[}wmuL==)\l3)ۍ5r//2&~ѐ(gv"ْ\^mtOiܺr$ 1ώ&PG=W̯}Ʒaq}.Zu\"\g0$vp}|Wx+:k7*SH1l^jT.XH*6 ^^?˦밚Z]6Y1)m8ǫn4E2sG?( ug^6z.MkWw"&>1J%Pךp;V%{zB! y,xR!Ǔ\]D[%_RC0 %p}\۟Ԍe&]$(Pd`1HkO&W9MMWt5sHMyFvA AN꼨Ǯ$tA!rg﫳*Ilru÷!tlT]f;o/;? \뗻AKMݠbCnPB}rxmNp3\̖qvYe39NL od.ݢ8l2m6xjms ލ#Adq>2{S=**͏ lT6T(?au d/޼5F|0.yؘ}Ơ9ԇ>}6s65ܷj3i`fesnF$NDhjq#eNirA1Y  TmSt|S-^gQg]U=uD0"$Fa%va;ƬǼcN̗pnAҪpDJw%H}m%:4̡ݗ LO9m̬*`%,;e6ZlI{E#u4գd}8-s Dbā'[Όq\7 *s6A|Pm{'[u_gR^>yE yk7ls'!%yREq_Q0ͫ r#`NV_z0Wi!9ߑ/Rz_{LtrU] <{ܰ<;_dqueVޟi=Iʂ]x=f}O|he^Y_8GgȀٰڄBv޶vp#MȾr]흊8u7(m!B@N½qcaJ.J(QWus.໿bv滯O;wiMg=|qasTՒ5$r ]*kݚA]?3S.P8|C[c!WHeصhlLgPqO"Cז[Ue 'NOS (d/ ʚkp~ٞ}pdECiuPQ-IN07D .luIG]giKYns( hn\'ۇqbJs0@P]?hDA=FK2?3Ѭ J%spns(vEl[V([b L''Ϡ0|tLXy_DbSVlR~a4x SFы8b0O C;yr5HL5RePXHwCax2%ϒӖ%,Iv8)}ƽLoqi0r*Y}Ph|Ib0?ퟏ<_?~=;rvK۾#gt?}0!;n/}眏O\7M|[޷C{-r_/Csd]̎v3 -Q!#!{Ukqp3gPf1lb+S߱*Z#h8PUl:ɾY1 ǀ&BIs(ъࣶ0FWW\tf q!Z":gG%?}3J\"'Kə'd~˃SxP]PɫH(Zġ #Gyg1"rģ ri C;?\cm"#") -Xi 8, U`IacIS!FTtEVtns AOΡ08~|VXbRA-*' #-䓷QfI [gPuԯ(_ _M_%_Vk^z_D0_pZ##p!k6v?Ca&7|OjH=\I%F9Fwc2׽P&>bŻ\~n&v~53G:(u}Cgq28^պSgP>;/A*$~[`kԉd;EwGmqq{~59RɰA!XQ[D%aCPu,fp9>|[|b=Sz}=P0y?gE CVٲ,IPs(lNT*k1*x6Kş 0RuF,!1-jfLJL= [sݦ7lo"uz`#,p*?,oF^hU[O}5u/~?a#GGs$`y} /*7^qgxìLڝJqsb%|VG8@U4=X.91GWLӥDq6EbٸmOetQ[I~92ceփfa?ðON.y|Pk`3JM&CXc*yj^5϶6] $;{: J|AӴ0|ar0YsR.,CKlJ%$KL"ko8ktmWԱ]~~s5Fd/J_XlCYawS12k^gxo-X1kه:F@x+$)s @qSŘ+m"LL5!;"BҨ,m)oX2v7$c1Σ8Bl΃-MąsokYY&{2vˡAvǔW-a Dַ+5U#ϫ;N1kU紧'^5< r9FFwblJ&{97o4,ND9:WګgU$UV[va"Ca~4|ֶV'^)2%9JprP?Zۋ[SQqͰ8T\%ZuJ9F4PAwUM`%62?]JһCL|q8bDq]&Ϡ0|DZ=kV>\jLES'90vZBbHN-jMgCaI)89[UεY7^/~1֐08JFb>s lMK)װL^b-DMg;AZ4Kk¡=|SuKtqGU;*Q,M`z5W(hI~k\$G䂓 bBfAx{{#'4;J~d ySyo37FFoWf'{*s ga@TS<%)V9'6J7/!yo28kfXK)(1nD웏!!R6:C !fk6s7OlΝR;j)ՕQ/vgrP;kѻϊe1'MBAϣT0P;d?I62ZXO9)d\!ʡ#TuN<"maأ ~L}cƩؕop"TO췬<޻ps'^.Nr:M\c{&h8<rgɂ W3/kWܩ9qfE3mt0WN3bWDr02T1],cL5 .* HCPDmyUo=5?Zx8şw^!cd6/VqP;)덚vNSz)3"'B3qXCR/i֗*(]*rva&Lʄ_|kL1還i=yB>VQs.LAOc -Q㙗3}? Z]o䂠_q7T_i<34w [oL.f0z*v$w~ G_N5]Q~|P9\{I+*5BK C2rQ0dla5;ʹ*,R*n! +X^=EQ-qLҜrE(dh[86JY_Hhu[GB{^$)P%xseH.JMeNsXTƘ'| ?F) vz򕎂e`4),7nyx}#A3-WCʃW9oBn\Os9|Mǣ4n !0]G鏛O1.6 ')YZnBέ-lslM\zZ꛸K1Tp!kNtx[<>l\ UN{ȟѸN&i gӗFG@wյ޶X:Lrx#?E>] ^B|__| v'e1):M]1enM]%?HK^JRrMMRfF5|Ynt{ ߹w3e8mݦVXQrp _/Kpla1eph _]82SLC Omˏ%Aᖩ4K:eM?*޳h"q'Z 5L1*  TO֐iu K5mHR6(*"a4BJa: D,%c$JRn #)ִ>ʲYG*\eUv y! [>t{{os8,kNP7(: A(}yk~pk6VͨDqq^ߥjb+\׸իv^~SVt*֯Ugnuw_>??߼ &[ Kj{oGlhs=7~4UlG-&ϕ9\#BsG'Үd J?m׿o\*xz7NŒv҆D޶e_aW5tT`Ϭư3e0 ֩5:7]@_[~}}5+e^Ā$~6j9<=} ݙU{O kOpT ڷ"KSݡ2':߭1k֯%xՀșyGP5m˱e;P3һǒ|jYlqx7n֮[8LgڝRz @aUa j. Қ$]^|l@l^N6tdu9Z1)nNZOT1yq&m'N]&:i߳p *C7w )83&R+rHim4t| .e tԽ? : Lm谶"J"jCP#; ZpR9mc#F#4r*D6gDZ#aricdF_=,{:uېo6 ?MYhc~vfmV9̼ h˅Ih%e~XwF {cu_*z'z&QQ[+a:eʫ:(uEoF1hZ#L_pV);<V;זu!VB'K7aj`USKeR2|ENO&3d&#QHYu 1k5f,`ZFL&Z iCyJ,\ܘ??ؿ_NUA+g,J CQ7/ qfo$*Jm8Y4K%~ҫT$z.d@Y #Pxt0a<WsEo 49,v˵%ʞ1iפ~~RR9_jvd>ZRHPG HS\-&xbe j0X&p94KQ wmQfRGd5exwv_5(Li F#L`J\B:`^el.D!F4ZqXBg 6(S;$0 rtmH葱>r` ab 5v0+Ce@l0Vo+@|7"g#@qo 9\ΗS6Ճ?@.:VqFOuspk=wZ o7X"絏{¢'6o~E"ZeAc,6ʌ ;6`#*$;ASن^oN],>L34-)*X]8N# N.8p8`swQIoaߑ#Li!bvN\-Z|V`ha)1H Цo`Ř:ha)*, [Nr0 \sc P' cILI YRĒE*L<)MSDcK)b΂wDðq$m8-EyUXʗ6(!/қfIlг6Cxn M5 PS):.4 % 190o/A;Mظ l5^F+K6Ӹc(R|%Srl#ǙwH!sD!(!OCnzV9MwMo͠Q[1"3\zԛXmxFU*F1w;#GEdR @1F*[HTEicM,j Q×XBCԉ[oZb/>"R8tDҀiW'bۧڐ2~^mBƵӴNx?>O5#iW^JjO,j@m4k4gzˎF/ovjNylVgq˻_~_~ʯoǟ~Lx߯?Gd Lא 0 |?5nݚkͭEQ}U!M񑍫oK~j}8 oRN;{RP|Ywг (MKқ^IK1L]LE3Q;OvQeJ۪^9OOO'˗*FJ=Z,ǿ[]㛈z0翯ȅm+}~տ_z7ǯ;AU:-괰WoB}f(W?޽5PK GQqhӱa˅kyN3򳧊՞ٴny{Ev/sa."d!&4$IL)r6+bd|L^@A1rAA==XI4~~'ل^|$ovZ͜ο[eZ o+!øcp'sWS="o|u>]2^N6ZWW2tugnby4L@[%IN4dJ+VkQJ{]gb=bZ|3n[nW.WvAFqե@hs׹׏z_2[p+l]?֛Nnt斤<}y6e-gŷvy[=/fn>@6u^jYז[g6W/n>yF~F_D򥀥.d[ĺR'hc#K].Rsx`Eb^A5ex&e-f7./}z2M2n7ɇ繳/e[?;rю3Akhs/va~]wv0eѕ)P쫹([- ̒.䦓AKϦV@;'==) eVVݺjo8e:| w'+$N0',NBɞG3LԼ|uaZۮ^%FԝҰn'p=V:.tfAiCr\_\ iBȲP}옦6 vX%Xj±a;.ϭ^3ۆ;&^\bܶwmsmTʤa620/m8Yw杲PiiD@'/ZIl^xɠI;-8)}ryZ(nT)4pdH0(wRz1(m$](Ϟ.0')1޸gRҦ l:x]u,d3 [qD`-h!X CO TRʻp01P" h'A9ͥ ! 0Y( KHEʁ3Ub̩%x"2f /:c8Gb ȝ "ך)yx*O|k9 d_~Mz^_ח!3^]P2Ӷ|Lc&#L3JeGc"CL3Tv3LRSQG]erAj~*Si:uՕRL.Aeowuǫe-+y{&k$ٌQaSৄJۻ?ƣi4l"|wµSfNy 6'bFG`NѨL.?5U_'Si稦AE͎H]iG\*+JS3TW(-YĹI̔?7z?lۢޚW/Rpdt2ZJhE{[ C @g6*H )MekeHaCT;*C e& oCht9F>+(aC#vPhxcRt&uvËw|w0!⍝$H% yߍZ)ݍJ}`H]awGTrzd~+f / \"ֳ80}+M.%p %;)T,{mb fr\1TP[l!TB1ASPY^^N3KýJ:)v+l8L~r20:csp`/̝$Rq]%!Uld ⃟T|UF胄 "#uF0]T;Hק\3U])\&Rx =p'Svm)TW1THKͰSYWʫ`+&vW"Z^9l%)a E[[KpUѫ?xd 3k.lD;HIւdNPVQx)u8y˥9ZЕI/;!DŽ}&$0`p1rA]H%ҀZZ΁4=ٝv4Y5io<^g j\AW̛odHjؙ}Wv:/{&q%C8/`Vx#+5%rֱ$|u'Lg< u&:^qP9X l$81IPDBgSjHks :7kqmZۃOs_=-5[w@oO@O AYqwbm;(.sȥCߦ] 2sĴ18s ɬڅ]~~};(wQVf`Pws?b1JHL451 o3THY"!(Oր1Y?1\yF{˘IyIB#ZQ1#4RBIrMl`*Vt" #x*Q(h~e(ʍbhE;N)}b.l\+yuՎ3ۡ('R|[Q*ialp#ONL?\ekلXE;=[T3JR ɤ!qPDq,tJª$R &TR5c1rn׌Q^ta1θ.º:]ʏOfFU\O4`02OO@cْ< 0fOwq5KZbDĄc"^65gQ`xJ%zqDsah2y SKQ !^9H.TT֖N:{B2f4%D*:$(J絈2J z@ ŭ%dl͆@'iuKI;j ہP3 iUԭVQl'tMf"* S8 5Q>0'׮/Fb(#aAZWq]%c5TWr+Q\ ?{^-*.0elL@DSz1. ,!1R&AJI 8lBg`2Qf$"@'ZnK1r ;'r~WqsKa׃Z_ByfJ>Tit*QKgxg"E;8[4XX+&)0JGMvh(Xj#VGjk%NҜ*yb Kp1rkmH_e6۱ :d\bSLR}R$!)qGHdq]_UWW8ŚwG_LB%ys b0V܏fגQpwA'r)9ZN={R6Wc%DunjZ'okJC6{= QbThvGyy9`ӗ`t峞6_co뻧yfC|5>9/n.1U;[nѻ),բZX{/,z,^>Q+I͙[Io+G%iKlǐDRyW.c[J9Ro}Lq/NJ+-R/DPUp$UwPǗj*zy=7 1j ei{욻% Gffhk:LIt|Y5'cކ ' F<)t! KK㰧(E^3qT"CΡ&l&]Z.6F7q(1aӯ6^V]9M!]zz ڏ|08E60gޱ<əh1p=wߦ#:$ꈎئ,VUpEȰqQ;d #ު$TP3#9,"xmd1Rq:E2.Jk#a &CViǬQFYJy>%lKȳ!xc8N%0qt` HOux0ehszGDJNcRɜRafXTs9$8"` (O XqbA½ P^,]~2:?*޳h"q'Z 5$1*  LRk~վiFaPmcp10"06 S^Fd-#5HO2s .U~4!C:U2K AvZ%)7fN%0A?ʟwa H7 [p\wF7c!/VTҬ0Ea]&[ .U L֏Y2l8F?z=T% >+_gss~w^.~yswWݛx#0 .!!1 ?oF&cq1niho4UlY|ve]i!2tnK;n" }헗=3t8n|pQqxbI$|fz>e40g]H(S1KDInƲᇾP5NYua[ORyvv:9~ĀAK' Vn$tw?'ݾ3/nE{ɗ o|s|عO,} Zg^|gc;_^%sY ȩQR p,lq/* .4XYa NsYMQzjZZSrR(g Ű~++ T0Lckmy=|jxFW,籷^''}*\8b&MwtuSpܙӤ]<; _aRZf\!f&-3ǂnn\b,8@ Om`m|QEtYdv5+xO}*5A{Хl "DYϩӈr-8 FKUvmì`^q^YqLY}TPl 0N.z˥MiX=jH蠼[OB gEOJGJFl\\z?iiOIڙse}]QlJc1F1sɭ`)$Ť<)q乵d42ZTYMQăN@ϹSD3E (N C Z4a#A2} wE%=]J1VنqRchN =2k3&TpN59jX|+4>cc>`Kphآ=#Aa[CGc.{0İ'4U,1HJ EmI 9Zr;'M-H$y:h7CXd;p&>c]Of-Elrw}!m&4} lĿ-E~syGԗskCWcI/:bL6w{ҙWb̓!ۺ7YPSrV`mtN-J[H'H%ǪqH{҄n2*9 &K7Ysɓ͹Q-`J86qQb63.2Ĥ*M9Dg#PكIʡvJqTQ&E_GEd(rbrX+"D46.1rV@B3XܒEmm8vFک5Gvd&,t>-Ԁd OOҳॉ S qUL:[nP.A0˯^^<;!ۀ'q4%uH(JS2(:o Y$"﵌FMFS7?g)w# kwܭ\g&5*\t/t.L fWT]WݿuVsD.% roLB[~P@bxLyYSaWtˆA>y07~:uX^\}+*&\1*[`?%/&Ksqu/yԇ@\7zwb b%>~ts875w؛6+7N/W2:e*<)np09s(Vpb#VV&.5 xk1٘ ZGK!zƌƁZYrV1b"*]j$h͆n6+8;7I;H= qR~@~ }>P}avYco޾L^P_($a~?i§Dl%ՐpR>?)f"'o3|wa(RA}mDr| v6#WO߿ވZLlF%9EAHV`t 0 >+r&` srV"W$&eaI-Ъ,/[mA o,l}β暕Uo~*DDfDna)|Ϡ1ed,yIdPv>Bǽ E1& @. =y)D#W@0hU"cWZI]\%*o(;0z~&{IVzNa<> E6/ JGjgr9S_߀F}N M wЌ eT0% .*30~/<0- >ܨ_} wӿvſ}/^GsA:?y-'S?kۈz:8J}qZPgSpaW,Mxtv#z2ٶHkOH s=e>EhN\ѱ(C?Q*Wxȋ!Ҹb|ns5P 0jrUN"AV y- qD],ygzn|hX)Q1waʽ3ø)a ")ДIz;2 gr|y^^N;[wmr'#=bYxri QZyOs:[;ӅOO> jQBB,[JL.0" yiPe NQaބ:`pA,3H\Gƒ*| Yh#,YydcN"k/%%\2N,s$k:Lfca g&FS8Qf^&y.Jlw ц_):$װh)"r):ڂ\aZn3l[:]짋ЧZaR3(vlpm=7Mߍtnp^y'8eV0}5{~r4J kΆp}k6{V_E]O15m|[9ujs->Բ?03LaAZ1E4<֛wbܬ7S |ɀ;7 sS5gj%7ih 5(o+= |&>x@1sm-IB;rO,t>tx:pCXhW;Ξı_)-U>|*$rKaEnOe<H9!g¶XoK_xc7iGdnjfalL/8}J0|=(7P vv*'G48NWcyZߛ ''yt6(,p!5 $C`&^19 (+@K"::dR6(l}XGc/2j98 Bb$#!q*XRE0I,tWڠuQ>)YSŚ8U^х&#U)qo}u;i<*~; ({<"NhRcMʂ!ΰJG%uS\ }>q1G ɉpj3{l iPĤNF{dh0bQe8(~eDL2A4#^+rHX([>mi "Cmz7|B>* ]L6|#ZnKl3uS)NgK3-qHP[kH<4q epFq >Kd]wRN2:>XdSϸJgxlDp+ )5p6SR=q,VuM=[zdוK&nAܦ(-vEgYȲGslRGpų)*ٱPf] 18'TЬuGe 4ɜ2T:mqʜi,S`hbWkuvu.$&Je9ƙ!&%4FgRZ:@1CR=xW^<%]ʤ%5JRº:;Kˣ=Kufib"PGMSRˉ;ǵ[1%΂4V˽]?08v"`֧ ,Jp%f7t]I]9r-hy3Хw:*B뻎~T)l:`:cl{X~qMϬ_͢ChB:XywI͝W=O\8tua97Fgȳ|Ci\ͺ5M7G6#ζyJ\kCg$<:P9GTy4o6%ܾ{e̙TЀSfJ3MXAD^!#ژ;T=dsQXP4Nҭh L Gvܐbqq1`mq_INwx=yi-e7f7S@=ob_FtJ]'}2NtS;L7Y-NNyRԓe':y[Z8'3}x.]tΗcݴefn+Γ?Žo[_;ǓqIr#_/a'>.3ö%LyD5NsJX7%y\, 1BfNZ1&LmO1w|T`LnE.Sx\Xm Q)hѱ ڃV>.8^cTuږaש-{%@dcj}qJGTڮ+`UzB^= mcC=}l,u.ZDQ2i FbO|lV鿵dn2K>o[uκec1X_..^yu/xE}a⭘ 鴺\wysvl?(!'UD;k :'jWMxM2N߾)X ̻0cGف*!y?  ܧO-se"DCv/p2Mj2R *sJkicˑ&ďޤ7hy4-lamԷm]7$yUd[U[S!ns3+.V r;ݍv/}+-KW^MKr4ɣ+Qs3:(K1rpXH3'%%`sJ`i0gYJfX[D 7.67Y39%Rx"%2dV0,(l$(k /( lpoo!%&`X{Α,˼Tgxs S:* DSBfPyP@bp3"0SLbHjEŒu.!m4QPTR [>$Jf(Q#N$!I8o$sLЇ%eUdq׍motA3r*eo1_.Up!`hc`IJ2ѱuOq2(_4Sk:ه䷲^|ݵ&߽ywח{|7 XQk ѝAe7vYO͇.14WMk24ksՂ7W9.(sziώ9^A@~H~r37ǞMW2-ӞV], etD #qc6jf6⇘~R*3nΦ1zI0Z$23[f<柳?'7o(Jp?_ Yi޻w,Erմgf!s.2Tw\7Mbbp&S^^e-&bנYkf(ɇl}zfڢZ6S|Z^YB1`epE!dy\Ov_@mٚ̋~)qYxV,jNeNbNLkVލAL,)Z}~^F3+M)hv"vق͜FXf%iܖ3 \L4{fu]ƮGoYdu5xjbrh"L, fB2A`WF$8NԎlUgS3)t&S*j5+-_qQYqL\TPV#2168å i%w.Dz鎄Z9:hpbhc6-v9:sG?!>in1i^E?&P;s^ż, f, `'![M_8)~_M;xXo/{m/30% - J.J. Dt$ RѼ=8EytmQxE.XÑnVMgxV-rmNwcٚ c8Mu+߾(DwRv~^RPoT6)Ĥ~^2s y¿ ۋeͨe1Ԣ9TN3ϋ\ː9re 3$)WB&{BiROW:5q.Qmnh@#EݧkOF.O/\ǧI2mXw]ܛOᝎi{'Ɠy/Ffh_I*gFmХ:wRD+ԺJyY,F.3p'sނdt0Kqʜid)Q;#ST~y-rv n"HL4r; 1) {I gXms1qLrD1 GA9N)זf;87"-NKgJ)1)$6:/=jkmy;,ogꈃ1*R .ES1bJl UQLGQv3ݴ0eiΗPF *삤m];g!\9ւfBljt)6K (G4U3f|tfib(F J)wk=ioGe/I!@I,N`Cb, )^d4:\*TL bt ΂بd,C0!B*[ [ D佖豉hj4B"FIɸ1p/is ؏2a?:~zq0ܵt'ޯteT|eLg+6 uJ>wIWr)he9]` :5x`?x<0Q|j,^9~ֻ]3=7Q*csr|zK]]V5p׍녔 ˯؛p_fVHmKF9tͮ['yӂbkZ8@ۊ:$B]sk G2+$>X ~6HTCd@9 mdSfC8aٳυDI~ BYsUL"%sx1a1S_"@Z放Q4o4 Fpк;<0Xة vBb[+rk-DTBdټ-?_eEؖ(eׁ1*̀e-3o1 o񠅛[ q&h\:U.ERSrS&m9b> ~/sGXa,Z`AG{Tf:=|qB`DC`gQ= vܩb1RRXƈRyWKd"R#̡MB8HnڃT p/Tj4vCP q`9-RLp7ܰ8aه]=-9zEׯ<2!1e iPc@&,FQdQT^9#|[*R DS:)+?) %N !"Q^He9\ i-΅HE*2rXERunK"2ZTGQ25 5o!\F'ō8F ,3.@}*1k$bHtL9 GP"QN܆98gX?tJ~8k_uzʠϑFZ3q̠ K&ٞro_Yoy$npU% &X&IJB+h}6yxOqrڱ*a3j`ADן^_Upȣu,އ킱 J0ث\WRGSJmΌ6>Dcr'4%Ma0vQٷ|+nqPH٢l(žЗGPy,) Wn@t9+})Q> D\2l&h: Z63ΰ9r#cFH 3/7N~7],}煳Z­#٬)TZ&w;#GEdR 7$PJѬS$*4W]UƢ2rdLh>B{ ñ#(RIUc `j0'Z)qBnLPot;"ESP:%h(fEwZtp0p<imz˛~xvuJb,G::C&@SG.hLFꩰ9Z?c +b`D`0 O.]DĐ4%kRi5m8<*rFH)`Np*Dy@;5ld~v zQx3bxp>뇿>z8w J>zt'l5´ p6H_ᦊI.ePg^U%~^GNZikY[\6o~͛W|~}yOxLͫ}uǗ< <b ѝAn3CqSCxáb ͚u+qgd4n!v܈w??~y;ilZ$o%%++b>ïpGmb.T VnPZ8$Y߇0PNY$C*pt=|R*8eQNSZ:I_Z}O#@Fc1]hE{5ʻҡm (/| oOA=l`)4iVT-1ΆR860Gxؠ<}};7rl1t0oQq,ɻWǢˏՃWujpڭT)ĀaY3 *mW O u pbCb9i=SyPÍ3">)K73WWe ")U = Drs}7OLLs!Ӹ6(@TlcrQȢ\83l?,ծ&fe~\r&h3֠^YA395xCN٠h)UjU*bV^O^qv,+'<ȩAiaJ.z˥%B>^ 5r%tRivÉ-ԯHm.C#c)R g5F:);H|^M4K.P0("(t@Hm>d;^O lrRcQ_j ?$Oua4aE`N6cc C3hġ̱#i$M:"Prg!iVHEH"kxU $,?Sx}oئ'L߆n{ ߇"& \9c֗{kC`wŲӝ m.Kh2yWFB UFjltBZJ[H"TTQt0,D'p ?-g 5idqY\'C>p!b2A-4A0ɵ r B3f4jҕsWL#,8ZI}.ĸg .6/I_." LVVWJfz$T)9vwX#9]+=+nF:{yV[W[vuufufJZMJZrMI &4x[1 o^!~~wWwUE?]L.CS.~z}qu!с\N)kDp_Jmc o^o-;ZTRkcV;KV!l1Jk<qEB3& Vc,is7yxLs;@]N-0%Qln%9 ,[k! 3Cs}&\h` 960j#;-$tU8tJӅ2Ґk2Fo8g$rh"^_e0UFUhab2*h%IS5e 8;`R6wv*}wO:Sǥt\"B1Ը򸂲*C [Uq?)R Q5{OU7oJ^Ujbk4Zcʰ+(OqZB `c-w Gip[i$aF5§r>כ{P 4& ?9mk: &=vjrhK;D=Lr 8AgCs&Ȱ, Ƹ#aREbD h$ӈv tsAF ӱ5qwGʙ6巂ޑ=#|H0]d|}L 2xW0ElFhSMPr^~ W[/fCXbJcZŞ䂫!}5T庾qhpBmT*(7@l澫 b''m]!}fNwXol)v5 :H' ECyx~a Q644׷`p)v:q1G]X UPcP(`$(wh*bު$e*.I ,!1RX39(.QB ?8DH,RD[Tm)FΆ"g}s5~ujr'6/}5ߥwmvˆ/5Dx[.z :rN%)">bSrGIvh(Xj#VG#$KڃNR"S%Ola.Fz~Rx2Ғg'_{=]B6 # ʧm\nWT@%7z?`AIr~*iuZu􈣙J]^cZHT]oV Y7+fެxoV Y7OÈoVfVjԾ7k*F;C<09 `(!4 EY(곕# UC+ד_B|U U!*W_&W U!*W$-m9f)]%Rmt4X@Jڣ%P>axe/Yv#v`rJDjJmYXj*[*LT U*UNlU@o1X 9y$>L.363 5m3-a?YVȊY#+~dŏ?GVȊY##l[ut6y۾o5,`B&sM27kh̍9ɁFSQ veBn2OR )_iffި;R)x&n DiFATb_]ޮ{|0Juvy5W rtj1|&=]v:r/҂OH=*W+ՓLkQE聴Q!'t- de3NZZ6S 5^rE4ōQ09e{|4{bKB6X8Pfe:V "c})7Cf8>;ŃK} 5YY`L_sA|s|8n[nI]Ϸ­V >_~~驤(RadY7G7;4A2~g >w/SXZ3,6Qr|ߕ3ʘ1ǘ,ʡK2Ş1Ty^0-қ^%|#YGji­'N2.tq@Yr${o@v؍#)|#5tؙ.BW+(ꏓ{f>L`eN[&ocz9̒[kgn;6a3{̱n^+C{bqvSVUZ83%}m}-"ޮ>x6Ӣaw[qR7 Kd8@#' QdPHQ G2'4Iҵx}y|]_ xxq,oզUGRIACkfa2 ?5}o|t#/F@3ZY+&#j<ļQV<~Dwh:ǜ}p|0YlyZVh FInIϥC-j_P:zI"OD@6eH29b a0e7JRQk"tDs) hacZ9T^OsEƫ߬8r .LɅǙ| eHˇh՗@7d!6Pƅi!tʤN%t{ufW7豹AE7x|h#*mxbI)r BNd7AD8QVxxeUԣ_%I35d^ *c%N+&,!GbGXn@>~~!!Ǔcry=kAyb|螯HjWLs`F$βF'>w`ٱqSǝv|t*/]g_\?çӏ~?~DAz_?}{<E8a/&/#%kh~͇8|h&C|qU)/1(^zo%^@}Gᇋiw2cnw|It1{u( +H6e|1'Bya5fi,Ɂ{= ۀ4:;?['ww tqoF>Ǜ{#QBoco}|Pr}/}ą^wL/E'U9gt's-h>;7P 4QFI`Oly'zZa| >ZOp~ZԶzv~Ѣzj8ڹŭۜ(˵j&44IV]ۏ[܍nPms6o'GvL,߳8_6'cn1꥽mP_m&ݻ.Me~.N }R˷bݻ6{ӐwӸ kиP۞ޒ (/(q:Yi]ݗC-_L0m)$:yjмrVDp'^h`t^H*&֞|s5i_qI GW\;BL2E\E^ Os>'K[-c !e~WBE*9؞29sDra/6&}u_O~PrBe2tʤer92iZiLZRZ&IFPQ>w{~푷?hvb|#,dʊyjպᜈEm'ZD(ԺiOħnSj@mzQg]U*^b~ f^w;[Rv/oQȎuMYCϡ:9DljŢ:5JNNHksFnqo >ps.5g bý=V,ZgZ&VqѝK`"KKN=0JmL[cX6YLai^8,0(7ZʟލGyģQ^| _y'飲,R:$H>&Aas_c25$t2L2qa22nB앤OO{5 9ɶRڒHPYz 2NFReGA4=0ZBVcf`)k]N)T[oZ`S"{&̩\1rrgS2̖wTŧ其P&QR8YtZ`&V䀁gYR6a4P K Hɖ\N*I$ Qt&\AMG/hk5&jg^QiXXDΒPPS[U1F Y tc=k&Ξzvz2&-oL` 5R" J,Fz4 *"t]Z[ C>6>'"$!A+ %bJcH16Q.A8~RВj16$,RW KVx}e7Jm0bdϗ59 GqR͉7R`3XG{Ub](KD%=4" ׂ*Ő$g!tmk#²26ONEM*m|uy SB}?6L S8qu5>3Ŷ##!-TBSDX)X<*sq^}R?`pF&gcJRڞ,vgߒ.2e%:)>\l&[5xoTփ#lxnmoN 촢lRPxi-XDE9⏱K ;   6EdI?: LnV0U}+q|dtm5ׯ%꧞/xH_2in g8_\NyU>?:ePDrI6% =2t+зCG&@%YjI<$RaW"P5dbHSy륶;P]BP uBN( :+oi,s3UtTљ*:SEL]s;SE׳Tљ*:SEgL3Ut'uTљ*:SEg֙*:SEgL3UtlO 2OHzNz IOzӤxZ I$ IԇNӉu:N':X;p 9!{/ãW_`+@!h+BWv&ueI)ɞ@m*Mރ4;F,gи,E6xo sҀJ kt 1]Eۘ`."T KZ5^"i U }T{˟m"cE,&ӷƯo>MdwG$X~LVgy]p*daMr8'5Ȑo%5ƒC0da/!Qg?.[ix)ӆ&;1htn{#7o` _hfZgdW<.y<յ"절^!QgQpݫyWk{5j^M 9@cY(|2%ҮR6ze 4=;3)dС<"/D"6>jǎ(,\i  `QR,* - ]ZfF)!$<>0y+ǻw δAW  <vs$%Ob|)ه (8m~M]ڗq`1nn= isòϖ~C/*s9[S TS1@KR!8 *1>Kw&ؚ䨈xӂD6~+M>64T;+w((s6WIJJ-R~*{`ǶnhN ~iY)o#R*ʘ!ޠ5=`R@#bt4EP@.ZRH~NE*g=,) yR6j&ΞaXp3yFr1yp3]cvXS}l9~Y˰|Mة^q'Wb)c L @ Jv%\*9%%M58(w6;)s^Sv&z,Cd/2jJUhL(zz<e#0OpxCO107ݳs?vǮzdtRxL"X**OQSbg2fI'Y]1h ž`הRn){O>TGWWl>^ Wjp컋CPXx))K%?[,q߳u96sJeY,bҮ <|~㪋?=EYe]oޜSxm PA36Է|;a?Q8=LݬN35xA>YRR|:\0m^-IKjk|v <^_Fm Wuz."[Sr{^\ GE h\ u|qߺql؁t9T ,"zmraq#frIG-R&i 㕿3nœN[XW!Z#\C妊& 7&ph]r_c 4w62iQl&Dr`F#sBrF\@(JJ3M Ѯ$CSlP0,M&$^XgZȑҘ/( LfI0.<毿b%K%R˖lz&NfM_EbNhdSss ".Vd $EahĽ tXTnW¨HF:r=歏6 Ei0S쑜\נFsmTxxl}v)xE#kX[me/gPq>;@|`Aрfrfjf*b nsza/{tovh1Y";K\ֺ8)FRDg*ġU,zrsu]D{hNj8iVgӔ*~X-R+,R3%m}R;?|sq{#WLR m4Jَ|Nd)0J[*:D/Bd6Ev]o7W&qfrkX.r.AYO˩͡ڈ8Q|h6Bd$傠Qމ|MP" jwL5LUßWNCzSt*ô`zˋ8޾~_~z2}ߜ'\<70 Inn=sGSգyGs <u7yr+{}s;^@yA7.3|5T1BA$WNX.7jO=ęﳽCA{Z鶑? 7mFr^׈mzY|7j)~_zw-zo.TiqzZn w_*^߼;q"*iPNCShq޾TO|6`5jïo5J2/ɖ;]_VGgQ{Ǫ/M˟&o-G¾vaqM) uڈ&4/86_mlqնy384#bA\Fd+Nq+|՘iʨ噷 gqPl3('<y4~~لV,D>6sz:δ, ow C^shACm{<'5 Ah_Q~OlwM#nby4L@[%IA'O WdVD [eO]4RhR:9,|35nCWxN )$%W-QHu^D:SpBjc)K9Hr;[[/wKmV.C#|zek>M&h2iI _BF_ q>ICqeEHZD.j[; SVM"#뾌78bQUZzg y2u2Gc 393:Փ"ur)"hhCL$]T_HZ2DU_/⠚`,byUUwKn9cO`u¹ie3|0od[01 _/Gtѥ:ViT* zD`&;8 X3RzvpR@}vpŗG\3tƢq;v5k%BRjLn?Y[1yiTZyGȤ(X@~ P\HR^yD$( THM{\Xύ$!\KT&HQ ψO956g>;tC?|l`<|T;~tkWP,v_i3㲞ȷb?E _ |~/>_ |~.>k>w 0R u.xQ<<rS uYQ'm89"xLGKq,@Z)(SiXz@ ] 2l/>_V;ZKRT |~/>jJRƉ>"x&3JR_\2-)>_ |~/""VC0>_K:ЁziijPNXR(TezRK^˃W/UPb.A<6k{1aTd֤j^*Fi.z %QMnpR&/K3r6̖ALFߛ3^^z۠^~_^fQU٬ :{oYgbM{kJp+f麕›w=4}[[JkMD}i"'=~d:[غn]O[_`|O6`t-klYun׽[=_7A -ׯpcdi4k{:6~"CnzMj|iΆ?ki B<}.SϻzhilsE%yDH hN3Sc8ruՙJV+ƀtZ=2XP>Nc 4"+_;p,LJB@aPoXcq,ɱXCv,>895lS?Z߯,/9k}o֗6["T\5οf zXP^a3`)k,8.DxP2#@'D̃WKo³Iwٸwmg˧s+m.'vztdG:./F~qwKosW~5,eo $PKH\jR d%Τ Q娀[nq(G:[ޗm="DWBr7wmmxL@,a.vUu& h_j_훛;?Y/xchQr?qҁoaSRH }$'u<3`hT]s'5Gte OPR\ R efR =/ܚosY]k0;7P\l㽜KHjF0k?F*ÉȀ!3sdZ-sdc9>;H),(!S$P}Ii ho#1 [KUђdb.Lq!G!Õ*0tT=KBIǨ.rb :#gC>kgf@5 5I?/mIF4'5%P<0:LPALP\In nd=/,% |y.ݨ7اx{ERJ\4"wmUV&4#Eƕ oeGEL~4gÎSѱ"HYJH-O p$@'J'#ZkqPi-=]ħ߯o"i4&+Q $* ^$ƻ6\<"H(9 'z|(ٖd'[a)h)Q8"qT\kAuV <,6jD R%Bz {tsel^D-wCD'8 Vok`i2xn3 H(A6Od΃M&~UCjl c\FӿWտ%4 ~2:_xϏO1Y[Vy+dtO 䓏 ;8[06f0!Egm݄mv̏m~Ɵz ca|V xe cӇ޹:j$F ̊hX$ J$wvr'ADjιYR m fVZf&h6˛$ }QLQ1L)ƔX1!C/ƔdczŘ4gT/CtS s|&d>4SC;1F 5IOmZ ;c$:'-uڕnCW,g$-\悠]LؼC\MZ;ٻ6W:{q>le1{C ؚCv/zXMRѮpKM(ć8[,MpzWhKKhgP.yg2>s@;u> :HZe)C *7}$ԃ? K+1Ws[zZ=|`w&Ywmޔ_^|}.oQ3^,Њ4XI(b1ju߱1XQ4815G+zss7D I_rRj0Vr7'ؕ}fU.VoN_uꩮ &8=FhI8dDc$PVV^դf5X !)DLmJ˾i&T xJ͕ /LOć`/JIR)6 H Y*a&qI\x gc!<˟NЊ|9'v\?\#QCGF%Ks~.9??aO %bjgNM]+G=]O^ ?OWȚ4M5Zc?Kdcˆ Y;LPzi[9}Qۙ~@ s>JkoQ[?lҡ^Wb1&c cX,['zi k {O#궯QO`??X5.s $a# ȾV׌I޹kΖF(PjsQ336QH!>?j\Yb@T 6bh48j6L:]#'uyV}\FpdܸE?ǣWkr_/nKoV;Ni9DcrR @%9g%xPVe#PhơsDz1&:L9&]ܲ-諵(HFalGǽ2J3\XXx'QjCu//r_`NO.~p\n IS^'^#c̮L@Lʳ⓾&$hc]t:SxmLaD0BK$JL.0(b('َxq˹04{AŃ-x0ʾlfN}/uÐ%TM:X9J2F/!o2d -(&Ie8*F57CJ9פ6\q~싈{ ₈#dkLLx[$&['!elMpZR#ێ:TXbi]͆Q/l Q; {Qݫv;ؐz}Nu!h{}ל"dީ/St(l\'' ;̉%@ex2f2h4yQ\ce Кl$"UA[BnXa6dhȁCb \ mU+\Jtf"\J984LӰVXizR:;7B|{뛝nΩOG }ΨMջ+Iӝ\5*D;b Ps…MZi­90 `RXKbm,jZ٢drYڔ }U2#n~;^]\\ɦ&w׎־9*[ 95+f9YeU̯_Q+*$}.ߩ|^IR0ʲTT{7>;|{ryfPտـWbWyrS۩;eixsWy+.gx5ck,-SBEiZ$gYYtS*R]i6x%Vh YeQ,:Z-ؠBhH& 5[2/g;_91=S{`ˏǹ>sӱmgmOO=uYuNBp!TQu:Ws_o&GtPp>Aj߯+DsT $ LI-"K~V;kt#^:z 1ٛ]s,c#SU CkgUZ™nsF׫_95\|Uj W\m*j*N484uMq%.Ӣ'5N;-ד%}.FC3,r) r}XG?k-CR(j57Z,{, M[bljN^2ko5uM w=};I?$\Q bW2fa_0/k=@ʬjrܦsgAB 4gВX m87aށsf8O3 Aٞ -7{}63f%(Rdw N0YdWHR2|0QI2XCT0f rbLq.9b@1y Es߆C2M?e^@NQ7$uAOh=e{mӃ|Wui.>7e\~[6zqי]*NOpUnM7Ż?S86[?S&!cBOnrʍ*,;r'jV}g$EaauM4,-C,.6mc|T#'Vm6'˜RQT r2,9ahgk0^}LQK.&0: {PE7Yb$X-}&fipWc eG-4U}d4LdgPFL\<vtI O3yҽ1GC0TQXtc֭gE /18.{3|x6m}xJ ۱z3VQq!\[DEDT+Prg\|k|ԚL؝ge5)4BQugSpTΦn1)bau nYjp7V;KKFuq̔#QndWJ6A8*.pNGSE-XUZoܫ軔n I}{; ]K|ۤV܈^<{RLΔRHKU"? c(`Sr `6Z >",Eo$9ЦƮbƥbL*!VNOWHޯJfW"O?;uױUװ˩jnD}B H+WdDc$.zu/W)ŋW'UcMSFXNIf`-W"}zQLΛk3LչȰXJDd_k$\-[cQeZ6lMc&I}(^5,1Mr6)0%i0&yeoWkayCv0=w 6 헁˘Gq-颱BUi)MAj 򒜳FEaD0BK$qPJb1裓QlGl8o\P{vڃ=ZPg`5/Y9SK0dI:U(@;16S?VIzEXPt2EuMB3S]36Lx&y:c,O悈c_D\1D|ï5&&Pn--3ؐ26yD8H^͵OïFvSs싋y0.\5dG}tyЎQwN?XY^KzkypJ޻M{7ñs3]J08;]nc=xUl%ٸ+N 'tS-6O&vSKZҒ%d,X?y^ȥ*P%f#iެ8~uni'QWWۄ6LUnQ#ɫ\jNq'Ep Q;n3wghAVAs|Z|׀eBD Ǚ-rFE#Cs?&?T-M~Z`4kP6K&RJŜ1pj^D] PTĕAH! 5Ak %gJZfqT;*E dZ]D8O_Sryr{Msvth ?JnF:9fsK:\$J1 S (O^*# wuNzA}}/WPPӶ-ơ8 `|ۙcdAkE?9 x)ENa*WHrUE ^٧ms=lo6"ZtvG iYdc yn%ڡ,'=$ XIkkEA.rJ R灥A[!D\ʏR7&Lp g\F"FS\І`Hh:%1pvID=6Y`^9e 2ٔ4N5B)0JD-%׈tCԧ [pE|]ڮw4$n*xUdZKrIV$0BzR-"BT\f BmW4[CP FëYu@Gh"D$1&*˭a9z9<IPW{DwsǷC]I&nR@(!ǐDRLQ. 2*o-0"5H*&gIIh !Ž(U J H;h.-!|u&~6=Vx'rQi1t3)F?7US,jKEW%T?EHQ S/KVޟIy?>e~xo_~ >=^?L}g׀ 2xOك{v5M7}5˫E7|&GQ6{o)n~~>nJ#n,)*0f^~F4TǽiPqE]4 ?mԫ]ȿzT9S2rKӨvo"R;[y{P_AKU}e~[տf_z7Ww}ǜUxC2ȧE~guwo*L˻\ `za7W(ώo]܎U攮&X\;gO5{+-Q~zQZ:ڥū[8kU5dM2_qޓ,n9wEh^.6&J\`/9y׭ɘٚ*Dd?xcrDey)U88s-}×k- `WjFJ@E0WsUk= L`p"X[d[HEQN`ziXךiTҔK9_hFkS4!$f,- Rm6Ksi=76t.i*,JzF5^~;Mo7kY,p<=5O>aK_ L,2[9^-S\xדrNܵ[Z3AFeDqm޶y mn2d:nX~8Mg]CtA]n>5kEZM7ě+{H^[(!H |ʟ($tg+Q Gl7([ ZR0sfv!fa:;2buHq`DyӚ++jdV ԦԽ; q&h\z):棣rmrwL{\\ BMe=ƺ-35Xtqo Ypl؞W:ףVzAצ!Wk6hb&gᘰ{\2I㕽1n:ŃN@4!tcz<4/63y`g@jqq6d8i6 J4}q z8I*t$!)wJyIc`$c1 =Ih*ϑ-0'ԨF%M4*yJŦӪVr.8 i#4jǞVy^T|3]UaЅg)qoU M64/4](\5D~3[(a43I nҗ2TDr7Hǂe(JӨ0 (O4!6( C}$JhhclF})W}ّqϵ*XN v QB(ǴHD'Y˃2X@Βt\# S37 29$ D NJ'53:GQ'UQ.I< n_JHD',RQyLEޠF4u8"EMw \{֕5vꚼBښO^m ́G2 NoXѓ)'S5̴oR3,⫅RBQc=reTUV򶓫 j?KrQXKW "ǕKfP_n=~{ MdB c(<ޟ ~yh8M; G9cR4¨lo?z 4L925T4BDtrc^G 6RJO\e5 ("*CI;!Wv$*6KIKcÏߞ/y{}E NȨXa` x/z܋ނ[ J,ĤSFI!'T#hF0 V(ɭ \R$FBC{bI) F!MP" jIb4lvJymBu^82k/WXWvKI5Wu$gg-%jzhsFy'%퍚ݩ[:]oMgasWqï}{|~}so(g74z|/7 ?x6C֭ߚRn5rk(r_U7Hmw ~0|ן4'fiw3<{Ϻ3^}b~9gByi6I!r}:oS-7ԨOh_>i[ՖdzyuQԣr~V;~QC}ϋAuB]556zMogU9JPW_7CA N^>)t@-5pխ>wGE?iy:ֽlt@Tg ({ػ^ o固ӋպQ.,.ճfqLI+&5`me<'owgb^?l̈W|=soz&c꥽:6o,?#<&Oby4~~'م^4o"9=cYy?y xP d>ۻ6Q0o(wl;]jͤbrYXЗrIRSLiE U6D=E#(sH̪j1+< )$%W-QHu^D:SpBjc)K%TdKr[{]/Kt߬:l =_Ua͇9a4cg?^vJi*@ʊ*c\Զrh[PAcNMXځAN#dGz9,cnswdf0Lf[#UZzg ydHefc 393I:ktK!&ji.*`8>d/SЭeНƱCGOQ!N1.9@jzm8K :D8\M!(/mBʘVM%Be˯^^=TzCeNH2Ǔyt mF{1aTd֤"Rhc"1M4Qt@ oDB0*QYíN*W( / gG>gNǟ F_z>o5';t8OҴ_']Apkf躵»w=}m΃:ȿ5m:^O^0Om?ۻNt,}}6e-vݝvy07DwYa"Z;:6^69*R!7zҜ=鮇`; W'O,XP^GP/{wF'nÓ W͚>6Zy8S(.+#@UĽQV4E]֟zȓ4m9=IC$<P΋gO;,hAq¢( U)\^ͱ G*B"ZB);es8U+P(I!F R¾s; ]}ykWj r^*N',PLTLyGr&# n8M;{ YTt ښ#] v/hB3)胕—= 7Wk>iH{y^ *(3ai)*'pJL09\ϤƅV*H_BZ%(at\JJ` ܙ+* +~F+BWz\W3̵9f<2>5?_] ׫Rj`;X= c( +9  +GI'GnOWўtG&GE))"QEj Xr(/ؤRтJF;RvJ9VxW}6LNy4.jg1)PV*Ny[zpezLA<]$]KbwGXm/wb^0KD6j_F*ÉȀ! sdZ-t[LrT%`@RЊ@2 7ixRHLAC0Rrʞ$Yg).H5*>02gIh :ENpvZ lvpBNdX8Ɉ&KM r)Ec&RN(Aq%*HN1("{މ@I\wO"Y)%.zp舶* ͈wqw6(ZŠ:?qish\@ RԒ p&VIN#Zk;QDq)bkbSˑ%0/C.fF>a¹2+.Zhxjn~k1F]T r8V9 ɱ_%7c[0(M80ڄD*GFMAJdRH|'c浅A:}C5A'ٰz;\:Tuy>w_>2#$td'"P`зC⾟Hѿp i)Ve8kdelHUNeS @u\g%^]Ho}%#{pZS 1PKwQb}1AxGU::VqVe e㣣.ec!'BZL*e iyM*;y CqQibW9XEtX@%z/n}~`ָA ›B(5 'oE.{ BPƹIB81 L$x(PĜS)O?{Wƍʔ\\*՝ڍ/I^.X"߯1CR#)Hc[3.Fr]ڡ`]z;_S4 '_LQZMFTgR}֦3}Uخ)j^ h?)z!A)5׈P o^]L.];7C5@Tme<&@ 1]vB} {KAmZ'S.ygҫ\[zeɴ32?r_`ҪټLyOa3Wɑ G͟ojaۣxne¡p·_6"PuP?}#/>~o6VlG6Rvb墡LT.5 -rf&?c5َUx9Z2W@\l{$t-moy6,_OὛ ?60A~VKoxվtrƅ7twÇό/Q U-T_"cWaۻ g.ݓ 2HE qyXM4?sגpYh~/'R\ǿ/ NֆZckkomY{ȃ-q9! W$"¤H9 VGgFIBA-4FDz/|$qDc4RlhD=2:pBFQN8e8{ťYn/Opy,5S;ӟ;e :sAmAѢmi˜Vw [=luV`|~]%`^?ym4lk)nMܫ|%6DZ@n̏?U2?%c8'x0Mi@ H+y rjn`|TUQhH$Ҙs ,QkXIΔt%T ut2b,Kؔ_vr74WFGnP{ϟZO˧ۖDQ2Z\i\|D%DAIЄPD[+!zwIo+ Nw,`T9oi2o7\ibm|Aa&W}>*ۇJۇpPh9#7pɵ{WH;퐩dW$uv OLbWZcvJA^#\k ~f*-2ADx0<0qLiصdk*J)Oޙ:&j0kg~KxLvZG y7/\{7B`ăaR;e[kHmuP}sr7#$;a 7540!~4Ʌfԏhǽ-_GM?&N Kǰ"gyLG48BUo1AEWq"S 8O|ʌW߶w[Dfpݟu:\2mcW)MuZ#|6Z0;>0hv*{WHY-t 3R9g5HU e_޾t>!01jbwːmƓ,Lݷ%xzxOd3@EE :1SrY/3@g BC>t m H[0_?dB[x}l:QB#j)` s,N qg'<'TBXg3y|[\{7;Nj+*eߘ?wK&lMحJvԼ6n࿽WN_-Ɣ%({4ڧ&WLZ]t.!`T.r#?yֲ)~H*7STZ덚k 㣞AaDYsÈԞ upa<ݾdюv6 udS^!ouZs+x.0;ri?m:ѹZ,VqZ|]|~d 6Jl=YIsèdЂ2#%F D.|Y<^7 dG*l Z]l`]It^hi0vAXIRTr L-IW%b%5)rU# {m$Z3ЌNT){%N-h"g+p0u h1Я?[~)򑗫+ݣL{ݟMVS!OJ!O0IkӍ!{sh{ȟK +8pZi=&%TԢk7M&M"LLHew`Ѽ%>%^A%L&wdj F*pb8H<&MHy@mEޜNsQUcIu-u8B:=*eS:IU |-KBqyǖuo創6di6>(u)Mb"!Ƥ9846HJFC)^oM6ht4Ih݄pnԞn>(!OV(}GRBbS1P o^UĎq:?rN)ͩ;I˔0*FFCw~1/`R>ZP16 5)u ם7wši]|dz{O4j>Z8f\P)/R|,HW"Fdu\"*m9* {? h-&y>:F/Orv"V)(VJpK 'mH+=D%WXB#q8C]"#CN+҄%,3ET%i1hN9+Y#W1ku!FO!KC㤦$gB#F͵+EB8ĢAJK PAĜR  (#OY0kGw U0>;I;i&#Z}?eTc1խ`߭]FπK2UBEB^8!MVEYS bX~"1ņ0a^ɓĭ2/e'^E m/׀kmF.{n7^`iQ`A@ΐJrR=,?d)2m+ $< o^e9GЎq)T22@-2l$E-z1\9LDgQYDocy"|͖Ⱦ [NCƻ,x%P$lXvAb8ml}鈸sy Ѯgyi= %^w~ڽkjrFp39.i;'~#zzX> =(#g!Qh4._(f`C"`cN%3iJ)A{h6H)!\Ixd ,Q"$iM&ԪD##6fnCc8H<+x?4z7F: vi$ˌr}J;3Un[TY ӄ dw Y뼋ShdaqǗ+ 9 _,Fӛm7,_*|T8Vrs ^7w#y&~c<ܚ̏-O^vq%O:3ٓ `:LK1N?9Ʉ#yb5Ԭ`d X}@b Czc@I D`d2C)M,@$JkQ&RX@HWLqXTeCEOKk`Ԭ;ہM-tH퇐6ށwS^[RvnkAA ?woGLЧfwTi̇]iP Vhqp+"WX3>g|s٢(bVR'˜Rd`4 EB2&W?:6@{ǹE;7<$zb yI9xJ)" uT_Zf)(ȌЖe# (KBQf[VWB$PF#3Jؘ5&=gp;›=ALtt*JO\yMuCelwNt7o0Zy$ҠJIȂb!(hudHгJgBΠ%ft Q{*Z*V==z TBR3RLLN:&VR5cl׌Ii2Յ6=GՅձ.uxeݞ3l8tq8Ng9006+ļʖ3r &y5AQ #,`mZ~ 1:VjS;@g3*{‡L&yѪ;5vh\n:ڍ~u`f_g椊ð UZQL+ 6Iw!9k}rmj3dF0Fu eNe1pdGTgӼug>I2d*CшǾqF4ֈͺCINBl}czE\=)傳-$8aA qVGrP*NcŃVǾ04ևa?}*luhцzY>ȝ*fY|~b!!}Qqpg hpȈ'QW !7>ƣlR~kqB SgJR] wŠ$q;.r=Y\X)b%DYbb)"cL) (P"geTcEK*C֠rET4*G$ #1rn֝C')bw>vd|r~щ-|KL 5kyO ;ȼO}EEE\FLF0O+%G(*9 HHJ&RJT5EI>r҅eDc()T<*_rJYt:P$ͺ]YOYT$v'+_|uhʻ9ŜanGfL̒'u=wjɜC_kr}~}sse\zt6|o%;'5xAV!+v}<]0VV2cDzxxxxðx0dk memL!W"b g2MxP $}:3%i1dLJ[!RB]4\&Kk_ql |H yǮƱmk_o1Rzv&`squ`~} #~zxaMaty7`vs_:5С*Echl ;l']Bm2̘@"etMɒ6;`T,m֝nCCK;Bâ2>Y, bJR G$$ա") 6tK .˽rE!DrR+}D$Pb,AW,I *@˽_rv#+S5Ba&-$Fڪ  Lijm`&kE9$ce>"琵/Mh1́d68x~ZmK+m+@[]M :&ĮdK3M[5^[]qyD}Y}_" xǰNJ6Imo4V([R( UV.'//c Bx/O824I+OR8I;mR8݋T^ـhp<-&gum & tգhVWNz z0iSo0ѻjf뭙 eP [/Wr020xΨjN5vy/y+Z i~L*t?ڿm{6]2l;ťͣY?Rm׶刯qaM0|ñy幗/o|iWV.˗v/׆;ᮆqF4x^pn\zxdV9 pZBG95ʑt5B *sk{ҵL1j%%>(~c&#fu16[ˮ/ mskUg.5z>e=$=wyJ R~dž\3H w\V7g<*[z6<=~`O|5\5?^oo<6M|n~2m0ԾbyOEnG07 y_,$M}h\_ԧn<ȽC[vU 9w"/Q_E:}'mzZ߽>^y)-l=}/=_7~z=h>tmHCŹp-:h?^>clȝޓ|q55mzhnc`ۼ?>]J6 i)7mڇhhc6Y{63^|=A>NAX-)CIyB qydؒ eOmX|E8E8ſxΕ?,%]xoxUG&.vZRr C"C>\<:>ʟ4C61&זWL\Rt\䛥jb14xsup\]p..|[-TǴȴܭ} Mϗv7>r+lsYvgtZm299OpH8zaÜjN&q,Bk!TTT R10@3"< "30 J V7)*17ݒS  W7&K,Xjlbt]$l5NbN- jN2X``6jF&#п9}?ӯK%au浻{[xLp su OW^yHb"me%ɴ ֧$Bާ€g \,.?ՙ}섞y8l%N LPiNQY]K)6oKC6y O빜yN AWF7o*UFSV)V\`17~`.N/1r~XoS6{(A:zhPNn 4Ļo }0^R*Q-Aw'm%v6/50Pu@&6yE*X3;Ǵ}1"xu~A/N_Ot?Ǯ陇Y oF=V},g81Y3i ʆ8 3mSϭ؄-dC Yv`Ŷ=wƒtbcmR9xN37fJDDr(KTIXP>SzNl$ZyCx6lpB qɨKgv[X1[( c Uk3s B>|JVfx2OKH pO5gNJ}T&leʖgmy$/M\I @ 4}ďҿ4,3&&qqRΉv+P==Jc$99UA$lN|%Hzs'~9Aݩפ9\aӭR}~nV?`+Q޾v9Q+#Y$fKDR2,8ݻw'_y7fJvLY>^l4n.3z%AYZ}: }=HUl+)P& &MQHL&RYDx12xtEc?GAu{Z$Vͤ,K NkQ,d[D<6] csԼoq2&9A]rDoAJ|;l8v =C#DOt}9SLVAF7˼tQoS_zmj(^utPBpb6F-EL` V5dFYsS"@d-waA< 8)Ɛ{'5Jf|Jd*T*)iR cX>e#sʅNf 6'Jq #7orlp]lgԻʼqtL@׍b eRX+z>Dᴚ&b| 4eo~_Ј1FBPuscC; g 7ka[n0U.f?iȮD W^HEq"pЅ5Z#D-˞cpU9x;#Mx:~Lۖux Bq6+ACofozRZ-<[۸ۗEz~c1aS_Af aaR *F,W:,Zbl㊌])u !l=n~Y݂dZ5} Z~ HM^)؄ lk.6*z6"_^,NNu!HݢfA}Bnb,_ɳLdߐ`X{3>%[WY-̺~<]^., /NPRs'ILP(M6%H=ٖҼL t 6686S4@Rg 2K%XwgVvjJ@Xn($.UKT" %|9q :1CIj26CV|L#.unϝo#.ױ2WEggS3©f}ky/=ey ǓF|SjdxYPO?~Z3YU5}4j>MTawu^Cz,kE/Va3Z"->;isW5oW?KN\~Y/~Xo-|Y~GD'm{78Jz=}veq5ϋj~G`]01B{_.xYMxeH'.OSK^SJ28{GO3B0_I-Rr1V=s`-z^,@s]1u]F=JJľB׍}ic~3krL 5J΃:~m !kuݱ9BշRf1ƻq/I@nVQ;2n6y9ݔ hX"hͿH=|grg\biD}4 J#dgvαI^ W/y5[ !5Tɧ^eJ%bkLHTnYr'2uއ\|RGH) (  ge#"aDe416]:&ǐv߈\s)r$VL}^sr?S7[ˉ/ E7O \bd|'mJvAfAL)27ƠwneYcuy!Wg3bDrFFI!?-bKV#A{u8TFŔ2P R+T3ŷ*1lO.>_w>Uzq[aO+]vjدE`8fcPR1Ar( :g%x.e5);"$%jRH(F*:Vz.VIjnD cF EAj428+3jf< usag«{ҙՉߖeo~_7g/ˆM-oEIU'*H \P^*T'mزg>)2SKPѥ$R41iiĨ<|q^oǂfcQ{pAĄ̈̄0dޗ>S4S(VݻVcı9*uRYoXS)5G%*qTrS01l8ʩ`u`<D6?@>6+oF+H>NuuֺlM\nYmjY9& D;%Lcw;Yc"ƭˣ'#I..c{W 0|BJ<ˤ,߯zf&E š5HfwOTW)MyЂ3.8"⣚Gvʧm_mK yO#NC\꒭U-.B˸=.x+PZeړc()DjEf*[A(ɭ tpI,hP=.;Zivx_] d?WgFu:Unuz=y?jŏ0'C ̯?M:I R ΐ&vȕEV NBi`VFÄ3C>sU]a^LOau7^Kq*=ܛU4 G{$Kòl>/Ds | 4q!o> F:jZ5o+Ԋr};^_Oc ,Tbs`CRa1rXD]ϋb8w릾߆"eMʼ83ߜ3_Z1:Jvqe˱,m8Y=Qw\Y:FWnzjO +hRᤎS) \Od q Rk TTpf =az\Xȥ\TҔK+"Zj^(Y\BH&IYZNkŭ\+F.W;o7'e6y}6 yvC.EUSPLI[ lY`!kԡd{@i9e]dJW;%т6lR@|@RhoJ'-)UZ`pִ׆4Po(ISL3,جk%b҂ @1kR>B۱m#qc:wTݜ›WݞcB[E/,]ynv{չO0yb>et}4,]btYzJen7 YU%K,9s"޼ۆZ͋wW(+ԼT3oTA>o7}=器Iy%n|rJtҜuV$ε&ԗeӜ&SoN܋_`=pe1ҫ9M]$<1IRJgXBM {MiF=JG8?xtvǻ| A` *1So 3ZTyuD3'ub RJ*h9Sp!ă6\0(3Y)<k1-[&ZgCY_G_URL[Tn=R@OOUn>o,z( W.Js&XC DKH4x&0&m,!z'D="Znq =0'HId ֳR&TXu7(J&=}= bTbe?76>qK*D qoAODu'աH%H%wQdg&^(XfP2m zd,Q$XUy5s( jõqP(I!F"t,102UmM ]Ll(Ggoڞ-Y$= Up|E~xX 77qK\h^FbBz(H1y"$k'BFAC¼`">tR3Csxr[E咛r*G/PA:v:V,J|$V@BBE~D\ dm2 $;7Pz_U |4i$S,]dh#PIA3r턎3iJpXiYȑI7gؙ>iʹh/J Dq1_Lͽ"z2y!Nqe1|1rϋX'Ϭ gσKF?˶ƞ+ \$;_qxm5R,>\53(ٰ>Ũ8sJ R2qq\|8~ͻwǯ~8 _^ 4~x\k^yӔ47kؤifߤ]f״{}: bnK~xcWf0_̼+#n-)}O;c(0a:`RByi4gXć3ll YF=/g4JFzYoTS⛈}6o՟gg#oϚ[+?V>_\/#"_^sV9DUbƂ#Sl| YTkuTsGVâ<7MZ ] x幧},*Kp_朗.5| N-.zq֭p+w;Ģq$Sۖ7m?UgpZ˥r]cu}2fާz~˝bdN訾V>:_IVL]wg Pn|4.8dۓ,Mqm\ CU:>CV[U_ ˊ≫XPrIQSUNHVăFI^z$[Ōf.+:~ (x-kbhbC⾨$SeZJPǽYuޭ3/"Cώjn쇆4~w4m'ⳛvhSt/[+A3g|6gY64hQʰQԕ5O~>8 g &j8w#lxEWjdv}X|]碞G_EC ٿzM1q@pr=Bq59BiˮJ =\=!8D P\*or(G%6Y1ɹܯADI~DUbx0jLgʚɑ_QӮc#qm>y^ɞԒzz(%HuWD Jf_4h@hӻQ~׊F++=.cCI4 YlQР«xūX'P#u}c<\|'iit4=29#?RMQ|cuH[.~h}o䞏^~fh8H6Lަ8>۴g z)WvV>GE՝ʲŸJ5䛿>/]JjNӽm[E@埠slJ7|hщqt༴i&->H!"ĖڝY,o\?XʻP}Z]2PgQe3ui+sTl1f#G60{pxd˺*,}栊>=NJk?փzUp/k!G?Ŋ;|m ͐㦅CzԺ8\kkYxmfumu>tG;VnpIdк7Lɹ ?O'!=ߏF?=ko:m.]7qoU(e5mŲR>ht&j 6pU"Q(s.{"6C r<w!2=xH̕WSg \氚1ϋCkw^JϱBɋ?z n$f@Eu'tҽ|cU_+Wг2aǿB'6eLM73aTؾ y iV\>FYRNyS'P]iMBHVԸ,51a0)c +w)%V^7+H =ʞZ5uv{jz/gۡ M)A}uVJD,S;p3/1Vћ9.DV6T]FEA+Bpzp"m [x7Һky7m{e>dTyymo{~s&~Sz-95~D Fy@alD24X91f5dC Ҥ2h+=-&!zĐ@y'"BJQ:"R,+JSE Y( yड,B0Wm̵jv^%yH0x3yWZ8wtK|ߜUj-?r8S?/9#^kY׏y&ɶmYd׎"Gm3;rv:' рgB h!餹"/"H2B>jKRg@ޛD:.Κq]^gUr(.ʸh{\qq'.nOBh@͔kAZI2h!k p'Z⡫0<<=ȵ:s7C_r*ܞʾh|G7T?>N4~]! &&C9;nXaG ^I^\}&yAeg(a|J"*%dqчD%bIJ[I2j "*CʒʒC$pZx[L#ޓj9 H1wR׷lI69zC![6'_zZz2J5uv[F?d2O>}}ߵv͜b0yE,,5)0}\}E[q}%P?4gZ'#2&ߙ>@ M}cutHw<x\Ά kgfedI rƠ X(1f"QLr-˗ߊ 3F< Iu5}({T2픉S GOFzwڊCPYLS!  N5:3,g{9:μ*$0 #Q clQyw9a /)S7QDvO&,g!z_t'%6+MEx"7;Lj_Ž.I,FoO98f;Q lȆFr)mz_I~[r?|k#r-ouu{=JfK;/O^U?Nj‹n[qX_ 9 <sʭ=s5s薹2ٖLJ!s}gg׃;S_n=3qe8Lii tZZvMDWz#L WY`7s刊6+AH/O#0eR(;>QK钶("K[0-zm~&qaj}ЏO-wW*\N!3TJ%28DLrNfX@xGA˸ @AY'(&Kg&99RJZ 2"!22ļ.m`tz،;Hf;"n;hnvNqsINX" mҩ`0E+)+Sx&M졧zz֐GţAG $UL ƣ% ;g.y Oq=ʺ^?T!hp Όꃱ*p(D cȳVqYbe#VN"H뭾?mFa]8ghRorj3aح*g%+K5@rktaw8::;*iC2GFdH 6Y0^ _nz'mTk<+.*E3d<dkzNEˈA.4=ë?E w엫qosϴ_m4NKCś=ă{3iR¥烜~gJH'I0LKzDcit4]"x9My9sLVgM/b?ٟtr{u=܇8}j ~4Il\] spLO6Mg"n^o0d^:,Q .ed4į/dů6Z% [^$ٓ{R-X_I 󡽂gWNj^RFbs?g'hebfY%'[QGWN8E#LKYRtL^+{$C= 1qTkhX&˔9G 2+o^;upQ4X܊aWs&=JPHShSAGd "EgqN@%A~kBiJS<™w%+隍mbKMP5Ȓ#X5k :1g[mfgWX%YJ&A)Blc 19Hm}* j6霕ٚ ʠT !cle$hhQFҙ#HmQz+? 2cS*M),;MQyN"K֧H%^O~IQbrxl[3Ͼn>UeJ=kj#''mB!vgo՚/Gǹ=tfrT'D>~/?~h{6?l$; 4܇wHOWwusg'Jx=XjB?׺-Xk#B,eQUkL1ou_~/No_(gMO` G:ڱݱ<|O(h_?~|ފ jrgG~ndpVi~Mm#<9; pkCgg{ Wզ^k{؊+si yONt.ݾm:o|l 7:˳ӮVk\>S輻nN)qN}|N,[ɽ̘2VZH] 3;;N,kˆnnn< hd)omΥ>:mJ6Xg'(p]h~+&.,U;;A&Ţ"m-L̚vXxK8$}v'ډĜp"ؘ}3xaSʵ$rVe.6پЃZ9Mrk1nv]^ 6Cwz9 jWmg?'GEr”|zS7uPEbc4D&q!㉝j{ _9V3nНVB޹g㟪iR*̦h\mNAU'P-({+!ڬ~'L1j%h"QVA>0.m (%&FIsk.׏>_|jit#d?amA058terܢ{BG 9%:ou۔2H?p>:7}QGvGmY*Z[><:p'\`oo u7/&mV;IV=/:l{ֵ2Ig][g_Ws0 h\h%#%Oǵ?; =>+ z;<>KӪXMei^e"m9]8-9\GK7^12{_or5C}3֏XnS}xhǓLpPxҦb=MNq_C wǓ.KCCߏOZt};/Gځ<;oU$ӝ`M A5wi*2{b?B89:#i<15 7}3SSNMa(@,R)-m`Arը]pVdPz \"9k}dB1핂Uxtt-[uG5hq5\Ikt=֯ߏ[e@tj[Ey[wOP/h/ŗTb?ĵ){~H)ö{(>ZB2Pvzjڬ3_d# LjDס#>-9[ :Wr#5wgS!%y~Cx:m/ɬ4e!+Χs|3@-{bc77/4 !ɹ+f={B-~0 vV}OoZY5Ly`*jlts8)pTW֊|ei.dQI"LlZ2jچR?g(|f4wA~ \{q'O:۟u!V(AV[a_Ɖ"iU$HFAH JY(xZYu}I.(!9 ʘ6E\L\7i|w0v)[$IYwFiMU1ȗv FI1AR%*u-2u[YZ!v![(um))CdI:I;,D > 0XtX Y)k* HMC)Fa]q]WX1`͵u tY!)!d*ҡaj;Ķ[Ұ&U`BE%S*DϿNYSf 4t,zWsh{P:>[ï{! V@ Z#D0 ڇ-;׻,1Kmk!X+O!E>%eZlm~A6 NXUb\DT,mQuZ WLQ ;pNJ?fǑ֑~]]"I@-I@c ;BԍŸzԼuQ\2CF*, v%d)1jD䒳"acV@jV YRBv.%ɊaH%CLe@dl0c k\Bnx$N2sVkO=0*<8<)v5$.[%liDcA}Py2WX"LU6!ad54նCFXgT-gۨg V4 6bPXAIi\ȆԔD*@*Ϟ\%rm]^NC5swc ['-)W*Zh*X4XIa2$3*ϐ6&pr8 .aLaǘQXj֍5$t;Anl 7wyΩ A0f c2o#1 ! X&P4R*T#m5CY,ޚrxgG5w p`Lv aar-I0xS` bXX£,"3P~ζ. mϠPg"!(.Hq‚UF8+Eze̯H_u0$R4u&b ]5udE\ll;kB>Zu $ki D$(Vc  e6i agmkkd[`4qȀgz(\a \ڥRZŬ䤁cQ%rvp')A/Pl]") ۹`Fх3MPdiUu0hKQ#Xe ͆)<"Y7,,9/3 va="4D i /}p 3 e0,qq@0ȋE%YZ{[xs;8:ƥ˄s"Y)GWϳW`;lC2dـq60)aaU0 v?s6Ajx1* 7D. .BJ -  BL)ڽDXRr,A4\`@;hر,ޗLA~qbZ9,RT.4HQ=#R*/s\<ٙ)oaol-ZÐ ;"PiQÃRŢ4Q5`zA@#dІEYq2B15R&H w 5XMg= |,|r\3)% @")I1?1XaojyK pMgAK* ZC .na75apJfH-tc/jD|e/1%8n#x3LJ,󦹧&3ҁ&`Jq7[K 8\4-iTgC1IU~y t٥`v Y%DMpd$| ]X.", &DQ"AC S !z cP a^XݤNokqt:@ykH!uH)e*&koK\?IV2;A;j Z%\CY;/D}m6IQ #$ދ2O sY~HEu*6Q?oϞZga1>,t &t1~oj b'!Ͳ*UHb{ǭ-Aahq`W%}+qs= I_Fu o yUMQ,Z&I=$ 7C)RI1P,vKߗgdgC%ɸeSqUB$iֱ$RkX(UE-Q3;kQΚs^+#)(*>XRkUV.\8[\ * 9'L&I&n^/~YVE0ƀ’) Kd.X &`)!m)>q Zi><{NJ07FQV+7 >NG_L i"BaF t^h8N`3x!w!4/^%Zd4$pˊE$2)zby"Y[ {ϖM/zv*LE5>֭tY0DȎ)Y"әEaB ϕA8k P˺dc!Z9 ,5AyQMP*p\vڡeQ{]ՙuۜl)1J}aDODODODODODODODODODODODODODODODODODODODODODODO_"9a%s-1>52r\F5`#TL@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! z@ :'$AB U@ %u_|H "$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B^$/J) @\ wH 5"\0B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$!^z[O,5Qe~YPw*%u4̏(&%~>%.֜ p ԁK bKtCI+ذqW \J+ͩ+RYrW]g@`ٸ+WsqW(:yX8J1cz=r>#w}6 ŵguHW鮬$0XuYԃj ep+.(/UÊ;wKbrPtB ~;&O?fp5/N4r莋}[Ŝ3)}{v\}ݶOu(wgA\UDQZmOM~nsk8;#wYqPZsQ%Jԁ^eI%UϽ#8q3S9>NZ˟]=NJN]G+EW8(0O._GeZ^7xM‡imYMo&X~_{L_QU|lզS,.zQ;SO^ ] ['[2~Ji>^˿aoo߀ya0+]a?y#{H1m(" p>Wa`YgSDL($k>?I_ |Kj3Jy͠-'8707Jo@s^ͅg{_W-KUST6n~(wף2Ng:%"O3\ZVgJ[^;y6w\Ǐc] w?^]b)G8xҳ{-GƣJn/8F[Ju}!:N~H0/]/opOf}ըcNt165gŕ(+bcU%Z.dū0†KVKJ+XYA2ŧ3wިHnEd*㦃=;k"il|fi|-")Y"әEaB^+pj$,8ux*6y3u0֥Sfrua_" '8ltS֑EmlDӼ?Ozlݿ̓Kׄ==W_Z<>p/J? HG-="PX)q6E(6R~I*}E\ƭ< M`FpY(D̻?-:)eT!ZP.;!mԛ=Xv`Eց%R"?Ӫlq4\Oz\ûmUU쎢Wm2c M׵H[4ښ JKw%w6.m]xxQK^s?2\hLs8cs铺ݫ!^zwdGߖaےߦܟc{r5p`uQ 0g%|<ޅ0TBJ|xVMv:8 ^z>?[c6)yx;ݼsYk#\yߖOwO>Sb~EZw^p_VZI۝cuE)Bg8Vl=%ؒ?? Tq= <+ AmĚCt })iϯ#iHvj:Q|2l91[EbbLjcb.%=i`hK ]c.:D2X1zT2Qmfj ^Aq8>JxeP4UW\i~ȂDž EME 2cV ['/qVxp /|?7`էwox** tsSRcCm&P߮(Rh>  HٌJZRu_}vt^Ù(伯 +lw]l;??~.)nS|H/k\5[5dB'j|3*y8[s &\OI.9msQPU[Fε TkL]J3qB/p-m~~u?pi8 x4p튇Y͐DEDAZߓ`V mUn NV=a6ɨhTg[!cW|QT*Kֵލh%~bnzƄ~-L*wrRaT[jU@RM_֑CzWFkMi[sZ drYtłɅ%#p@> >XbYКf8aԯj1Glz$’G$[ !ځCyEwmHޡcX|psv  Y[IJߧbm[bm9i ,"MVWVImC / WXcԭ;)Ⱥq֨/JET9)&5տ*ZZτK#$RGf\dlW.b/oEƚ {0hĢFh B`V@UH"9([gzr`pbWw,#Uj/|֞{e=x7'5[yw>(ڀ=g~>1PW>ƣlR>w2!6*`/]#͉dHa ETTfPDUjDbEK":g-*KmR:"b"gQlWBTdX>`QEIUY^!"98#ap!03 rCj'ϗ_b>'O "Olߝ'Y&|l\ ~JP #P3:.oRJNb4d" ArpbIQ4)K2:J]P:#ng}|<geW w/?%+7s s^kϣYʥݕfيblryԮM ӟuB >M>5ooƣYjAfx, iNtBP!IttF+d隃-ADWJS2,uBt5ȻgtGR'$C>V"(Vi}$ )f㙸LF]3]s7]Cv/|Tivőz.R)˄ O0 knfkx*/8@xez`MƧy,:"g돧Jt.7!6_qq.5JI*XV\76"(X\QC42x-m=O2̘@I'Ά2hoF^댜=|!!q9E( O)C |/ZGmaHhU02J/$<݅PeL"h&(IWBс]D$8[B(+$^@ErN#- )"%$FJCa5lYJԛnqͥIN`׾i]EwjԛNyA`ب|襠+/tVC/]4/ rs^au).]*DXԫ &#`JhJ>jom|bh#r4%#2ՙ۵=_{DAj|7䲁h͵>ڵPn|7-YK?'Q]{iywg'njT7 'Ն\; U?7|%y hb$O3еctɰmlN6fL8.38zkNޯ4gԳp!ٜL*Vc-{:~U^!S+f^^lMa8vo8F׽6k3GˢC|}ImVj[ ܠa2^S^o\J&xb]*#^EY1VRE&\hk{mKմh8u.YtQ*ZeE9a7!ᅉ1Z˙J%ԉKhg%og{|E{!#|s=U65'XyrQVk~hsݧFY J٠3bi|pAم&tJB膀:W֓'x}̺OS猤Zgj MCZ'|N)c~mlCPIVDX-q^,dEl/5HQT!ge]=\H_Iֵn;CN Slڝv}ͳ0M$i|tg',6w=:z&td 1o@, ֚ T,EcS CJP&9 <`=HzHG>Oà;{a⹟=nwhKCO-Rٛ .Io|J@yP}<$3nW [[̫`[9ǿjصuh|1aܩbyOEoG7`ɼ/-eu$M.:u6]6ԆjPw`ct Nz qٶ+HE[7ܺm[=?NWm!6?ͶCnpf[{u;_(+{PKo=sg3qzGNjs4ȑWf}\Nt+7w0^yJ3ҡ-$|htziGv...:L‡m!6CaM]>({M4uupٟuXUTM&"9k},=U6 m^ )"%£H۸ `x fR!$(9zg3EIXd١r"8[ -e\IQ+LU9.JLIA (Cgq?ndV%6%4e$N[+;Usb[&5c>OMQ;[&Xb d]Acmvb Nd㘥Ry]˲HE/xAxҽDTiW."8zf>Z Q2P> hH@pZڐ\!UF{PY}qAV;ouZ}G"PԵf7w ٟȎRB2.EW.ƍh5$UYTCqi29x) RZU:CXOr1~dIW ?U! SZ토01%ޡrꖜ^ZczTB'my ^K+T[- x~.NIjoͯ:- Zp{=*W?׫l?'-5gy;1]]ѻ':Gf͍ {?2Ӯwh^=޴ivuݥեU\Nsey!Ye_W33b$BMn~ 0f .ȭ!b, )V~}g"jD$R%NOub?O4F w১K65ó nxI9pQi>v dIH \O ͆[(eC7=[Û_U= Ż&V}UT9 X'υ)Fu,RZXZDS ')yΜ}WR2qcoNelouR>o[-9j T$پ+VҝgWKN{vٕ₋}bWxq+vѕt_b[2X2ֳȮ+j|>Ey/6>([j"+>u0WM&j~G(aǭZuq>>)J>IpVݸLp?0IlX)t=Yp FpW9[ l%/KRC)Gqi]t),g`u\ԗc PjJyܖĕ_oNre)YRHl qZYt]&ƋUns**_姣&{&(AsH(Z3+a܊-/;whU /6ZpT \(+ox$Dvs{I{5k{7$3ʄ4IؗICUJ{e D lө')DbAZe@ bԚ GtDs)X34F; w}]A^f9Ǒ}><L5ʼnkЭ]Z  \x:QKǔDI&‰1O [B( r=jIҠ@B/J@y %jDŽhnMƞCȗ|dZp:+-&e,|xp2>cn$ՠJc"~`xģ:, ~JGbTg4v!=iwVGJe{l<8>}˧?/G?}>L}Gy/85 (p[vl3wh5]u-tͻu=mU]߻# e\:$ /\|? Fxs*˙,,)Hq1 $ AUh: 8?W ,ߧ6`G< C'ԭʓpx~z6\`#գΆokLj?o~~iN:7 Y򥺭\?_7w߾e*.Q% x%,el`6>_Eۻ;0ʵFIzVŖKNWc h!/(jKqŇU(c[hQ*S;*Vy4|e<W-}+~A`伊+b7tj'ϹÛd*B=gG拜A1ɠIq,=<8sI.Դ"Y\|7K6sxxnq1\4ˠqá=Fy?YMdvu Rmr ˣ`R.IuԠz$SZ]` '^hhR:9UgU,Ĭ$E{Ŏ'w\8ډ %iH| N(ZB_ ur%S;&_"Q\ 堯6oU8kjTē__B\7bt:peE<j]pNDᢶ\p-4ۤsk zd%y_w&rrn(*-3 fuFsܦ Bx|X RzFNHLTK!&jig⇚ >i XN iAO]B)S߮jw;ﳱx>w Yv]3_oT͇c/td lgk!M*^Hz~:{ލ ?/QFvg #\#TCءQ n2)ʴ6.r,'\)@HRhGNNM @ԴۋI@X@.+&q*OxD ψO9@:xg̣Zh !}:\KfMޡթxb1J&ih;NoG D!jN*WFt)^vNKΊ7rr>* ɓq[k%!&p=)47n!`c dno~2e?+(\޵ƑEЧp(I/rݷ >䝑߯1d#^dȮǯzڇ-No7Z@޾ ;whC=?F χCZr*n'ſ%V4-Pmç&aҭkGr~n^8#]2煨..<0I$e؆ *Uo}~lHp.$|2pyw._9Е+Ȯ+"[rvz΁I$b9-)_.;*!bN"8>Y{[fNHtQqel`8E4ڎj(a:e'4ߖb!ɬEAr&HF)* Rca#W90nP-ϚƎ-SZ<ۧgC\8x]K$rjcoT0k`-HX.&tXc@,`MM4 W9OYS1e%{"EGֱ!c,%HO=zQ$sR&1;/< dё,0֗)zQsTfoT.Qm7 -n4~L ]3ai}#e!/ޣo`zs]t J,FlZ˃ {d *B\._r $oB0"R ֜)mY'Nk5!%ٱ`]ˊr5:W~>fWx/5ȘqJ2 d#C&,%Pj ٽL'˶B/y;ߵ2G({soߠEwu2͡eurvrŴ/Jrs?*a ]\LY_wȖ mw=Gz wG^V3ŗX٣y"9jz>+?姣I:j'{s׋H^-p¿sW e5v.LXӴɹy"ښcDE}aS-ا^O"e ӐD&촯=.B; ˑ@ruqH I@Yc- 0I$d ]**Fʢ㚑w[ ɡCzpgJ|:mi{w~#8zjSsNIh}>^M. κz. (H?qD?`utT_GىC%9 -lZbB1KT e`S`hpL/RG&+F/H""u2D ča3r+=ovA@{Q垼KA) &A#Tjz$%΋4!+l[XGx%YOdLA%" UfC3MuJj,"oM2~gauP )d(w~tqI,Zk=Q~]tc|YWl..nSE^6QȒ\%[5%tJb4H>:drm! :[r.LL-墥bC%gl.>(R X(Q) JflF㚱Vi׊3vՅƕur)^^Ud!@e~yb0,REUW(v";J61." /me2#C E'MuM&Pr 8Fgة&ZhFw!U1E#6]5b"ʎqԈmdqղ2R< [o7uvI{Uo?0>q 1E?z):qrqZCUP5Gs,;3Uk7ׅpyf'Zg6` CgD]tHB&>hG4Ъj MO}4=} ):KuZGڔ\{##xwfyŧ$S6|'w2!?=~ݼ:y*$2g{hZΞOTxVTb}e2%Y6&F=M @i` ށr7ݨ{T=yY[$`F5La1V׼:}8E㛡 3}1jT_0{_C_#]rh×6yy7woۋ= q@nyߎwKnvf=f, <Nj:ٔ}}/  'H;(QuK #HEg)e9&(@;`'۴WiDRh pX ;Syr BbKU**ڱ` PLlE**4I&B`*R5#gˤ)R=Oӳ "lz}3t~S5k5rGg$.U1ij#LJJ+RfJu*ڦXSbH^J>@#oRpH22WP*eBUhЌ%^xzqqY65X{uL<{O1_WfVJVw%r7˿ \HKʓұ+ާcfѮ-*$?V9 B\H7  ÅBgCdʘKvJ{1JDWHOiFp6u!} gj8.K| YJKc&Gv0@@QF]Z:"qC;Ǝ^~ktÎcH]6,=mu!LF0&W]1#^FxU+k J͚]LhcX-J:W4K-ҝbCۑm./&P[ˆf_TWs(ir0 :/IG!t)@فY]KM/ڨ-[" n|0P6Yw7}b{laGg V`MJ1ϥAv@[T*[|$V3O[441 ؙlR2?^!7ƂXX6ھ(oXiC;4Eg%gH}+,lTVHR[^&t*)e &E/W*/"Z6٢'mt{ܚycrV 6гԱMySV3Ѧ}olH3ۤ]Nbo*7uFquJOF(B1Io.H7'-zn0ƽQl\hi|ί_!xor;FiٺySe28>|ԇ }CzIEgp!hM 2bQ:jsAy#{29I91 OĆe2!2lOl#z#Q@X e,2H R)c@X DiN}(gk;O=w :CP> P5ihQ*HbpI=Q+Ԍ )0#JeBVE&C*ll%} P Xb :2mԝ5%vP t"әun&8m͝?BlzXYByճ뛯iϯMf?BE41j  *9 9/B˾SNV l u;yw>H 6>Cm`I<bǯᷣy#}#oz`]ܶ3DK0r= pK@&;@Ry)AU;{RƣR5Z_[૩ʅ2RE6IeUJ"f8"?DsX(BJU2VpdeAĨtv(tJ Ɛ6^ f2onw|=~_N1Hг-7jYtQ?ۣzS9+ ]S Y.}ϑ:O'0ᠹ㸛b+gœj 2)eu)L6P+J Twrhd/fVQmU:՜Me*Ht&ݦgϭ}BTq' _?O|$Ɨ(PTQ\lJBk;uJJ"Dh}\Fk/P#Wn_qI4$hlK1gLΎm:{"O,^h|y_O*#:90#)+鑑xSכMPҠǷ]ʚޘ#(/[jOgߋ{di@&g+NёFW#Y&b1U&ZZmz FT8 {mARtv#c?/;8 ;1n£bÌڵ5f峯3Zv_4\\~+#v(QUH ܬ(W[`.9B&YD -'k UNI8[W[AnC{͹5شCh"cmBb+C?Vǡݹ̡OVFYlʂCMj{ ˥b5MU#.*XE ٗ1yl~:BfAT4ق5\ySL /Q0JZ+Q]03m:p.W1X OCsCFDYPh].bՅ1A)J,Z|;TMu}"0=A+DVQ[F0LLZ""Њ6I;_t(QgٰtS:JÈ#.ne}=w*^94J{#D2F+1'{C0vxNnȵ:-{^%O1)0yApcE? FA˄3t.6So- c;蟰Drj/Ij<wE)6Sf3à*&,z>RԶ=rr*'gO ͩT!k+ES8٪vRv1IRTž l{ Z2TEG V$0%[t`IdMgϨiN4VR@~9o/{~2ewÆuw虽CXyR55yʑa#+KerfGrrKʘZ"FEfMƮL5/QkcҩXHDG)Xsѹ2qmz?6g}?^_M-W^{sru^S,6 08+K-ʐKICEN܂ bp^5) zIգɫG_QbR +PжHX ܤ#s5Ia@ʠrϟ ?~lknlhMv=N?eY?N>FFUxG4s_:TS!]R/N2H/LJ#<ه)jLv YDK RC;@&Z( K.)!q l$x0܅7/>q[6kMݸԈvoaLCPmos1H=bvr;f;>ٮV&yݜ\RovƗV{X~2rzh6[5JK; /_U?Nj‹ai.5t؅Pc^S?M£moͭ|k'*b}R_TLVP)9*J]l(R94WS/?٧/Ji WN'2ug7*yӍz ߃#LW`)FEvfCvȲ$:Hvc1&)QXem"`Z>Xo~dNo[c}yFrO>K:yDMPe;Nr\r~B\Oo/x3"GCc栳&bU0ȦBl|Td7"Oit=ѤFZgo3ZI$8*ET ÷EK`8z}7|v_у842kREz+˼{'ى/Aߚqgƽ!Xԡ8A~YvK9+fsyZHFM6rMY؀̴Bd&1z}NwNk6?x}5;n_[쉞'ճ hvN~HWo(jv/=;$¯dM7YDEd8$n2̓iܡZ}i7᥹Lf#[ſq«*wfN?:˿G;{u~v~FS>0;g.\"zOpd^7p+l4\BvEg'mؽ}G%e2|y! -)o3b\jcaSx_[xR~ޱw- *>+T>Yw4yί8S_k| ~ZctY OpeE I+z4ҡ6E ڜۋ{Ŋѻ~CwE)3]xn|Y^hhܤP䢸`.NJz9xYIE^-;A+4ircf/1dNF'i+߽v?9mq0=zs+V/7~>h~)6`EU*#dRSL*Q4}L>ȗZg.Ag'>^g[??>%bd=c"VU_ˊ#8mta S*("5Fwr{ws:U;C BfpxIˎx!Zo .ިЅULB/^ Kg?,p0RC@R68faO0:J8N%3\ *hag`ObnZy`ʆ&a, v2oU# SТ2>a3DMURBPٻFUϳ!`ld;k6ˆ%f)$+A{!#Z=iVW_U"'^qބ4!<1eţh 4e)qQXiϱ3}5iͲ,Q4נ|eb/ cruvP^]4=As 9XQYd&)0񊃋}?D̙l:G8xHПKbA8segdzN]?wg瘨;=?6`a!7 >s?y5WwMk5ksՄ7W95><>,[R[ҞIQ)moRb 'n]{R^ >3+YYZq*+ya6I+V&]ӹ qh7\>'u{|һW,=͊yj.e'|.ກ\h`xbP5o׋ӱj\UPaϭJ$xڿ-?-?\O/N-Ua . lEYvWtT̗a$Ay10#|A]#^u<_7&gc*;{8(vTUKƙ2->9IJ4@3e+yei߯ˠ`ڞ~fp {z XN6ZWR`bI&+7nb90[Fk0 J"άHi9eT)nx!3WQì8>sOᴌ[d#JdCpIq*mx$u[.m(1P+*u=ck6R7w SЗ5, jWuO/$5 G?ՋR+\Bks~1-CQix`,(3xn23Ńt׀"vhLȽGZTYMQăM@pJs)kfR!A[8)`k U !fH K)Oc.*5t (Xe[I^{j6 Rs޽wZ]}1,(.`1WkЃmVءBGb1CGU4^JI㢢oYfz)Zwߕߩw*G390ySc'g!lϿ O?~Jցqч)3^dLp8KnN|zwRbwT` >(ფh?Pi6NgNhto[?IM7u< TA!*i7IV/pld%fEoPwP/T4;`R  ;uܗdv=>`7fwxxryU_K/.yc4h=W $s,G6l}3|%gS 3dBF{!#M[m<>6RnZ*Wg6Wː5 X }_eaM]B }(\GYmBwh-}6>57{exĂUf0n]nBm墄3OlC-˅b-%o*,PWd&z5`bW Ĥߠ%ւ"qS~5*KkW[]\%.5WBr+W0y8ؖJ2ZUVCWWoP\I(H\%AW.#赈-?xc0q)y'ޢRJ3-~N;}VΫtWWfA!~ TopZ60}X;<ǤH$'puEM o:BS0LqxY D*s&-7(˔\A[/Ncd6T$80 MB:NT{n XSZ2(:oAZjX$"﵌FMFSW /[caAR6 r0N5%@}t^&LV@^T7 8 ZM'fVv/h&\-+-ܜp4-h¬Yc&76ϡLBI'؀?.& >Ρu1:/[o&9l!:k̞<AhC˼ڭn獷P^hy? ꧛4ɞb1W&\\džf[~ݴ-106imt}^Id/ZnmBbXoo ~-N\F޶`3C2r3dln5`#6TR"KRtM}~}=m3x~3Ay`a&Zr 88* T(˥cK9V=K*LjdA+냉KM-a$EV 1=31B)bKqoB`q4mFϽtq>T k]ML swh훟֡yl !s5( o"ĵօ|3TrY-"_ 7E^+[4J^Wp-5⬱#d9+IY\?Ւށ -٨ԇԸY"l.+Y5ۋYQ73#Q)6nX*R<~Y+;1!Ӡm.8*\ޕ{[l Go>9J5Ňv.+ЅU|FYLY`m>,U\g>,UZ?,,zkb<4^{ޚ0oseB&%3].Rt-z2pc#ƍ(kql} ~fqy{WqM|zk鍎 C'?r2_:i}x>,O;39)'@ m ' ;'!&}oO@*ߞ<{ϟN)s>\#‚w+~i'o^Eko6)ڴ7)ZVyM (,^@Z?|)~s4>(DsVGvN'|e2DGI^|JA(9F'E;MI>QG+s뽋:% k={&y7z%T7xUox8R˾φr=p3=z|Hen~;_%sh_.{DA3'uux@uJ. yzNpJqUgsK0 ivpL@# jo4<“UI}F4Caƣ:5|H&/۞奚i_Sm/L_~~N!4( gX)%"S>~#4.=Q7yrofWvUKpx}7xp+?ܭ87r;ֻl>Òw{|Vf/! R8i|2wšVG"DYM1Jp<|a}%z1.Wu26Id r /\J`J1;"Re#dGhT-X_nqz[vm yW˰jG4[椟eq2VXt dm2'zҽhQv$MLK6Q ȯg[on{{W҃>ZFˆ 9W6>kd直Ggl_;I~#h*u.`-35G=1F-84&I:4HӥHkFo( kr9pƊ~|H#{+hRWCNN:E9 'os#}_C2}9=y;]iW-,q']o$m<5z7On'^?'}ube1..ogWCBDCgԻ'g2a ãbM|%tA {E(c:"\"/2vw5HRfa  eщ([N!_6x& H*&v@ [36`x6:ualwH랐Sn$og8`SsSh\gh2]Yu-(KbܕBfBfa )yKkhQ$g#d$cb|mH!AvTI]T')$)E&H ɚd- c=VܼGZmv} /GeA*9,Q51qx`CAHGc"'H<>LʃNsinǨfV;=3R BRh`%"vNMEf4/$d@R*F UY[OO޹DZĠY3q6ԳQ7 T=y)iR &] R}a NG"E`: T7__ }As<%UF<%#T0ȉ9폎"e*dZ͠V>=I;? d"m Yy'zrqcB1<:KOB/kp41b7_}=7FD: ! 1.E @ MFb5" 9!JI;mk"²ճK5$w|,ll|bm`/(gjjNE00 pXvd;h+0ƢFȶEгyXs'M{ݷWcO]K?/UdWwsuVy;d%{f D.,VBfTSgBd(B pcR쳷%$׈9@ӱgTkL 6*Zmma=$ewNmJZӳ77/5|}tQuK؞V>fىI`WQ,3;U?c`-u1 c $tJuͦvI Pط>eJ"287[t2 .Xvѱnju#>H`KzbT1b@QU+ M;ʐ>9j˝/A3tb4EgMIdjRajXm"SM[o97f{;5ƱXfc[xY",Yr4CZ="$C4IH*R ɺ)UlEAWvK!ԴTӨ@jJ35u7دM>g"ʀfugn87[ķ].λri(.v.^ {Sg)ZHp‚)V㬎Y#AB8xwaG[mahlvLXk8C~GLp\!84 n~'~Hس+ljc |JeODdMvدƒGZG~+OYᦨ٥W'wݽ?> ft+qÅw߾?~Un]F޼F3vl})Dmug=$Ί{f߃_  _# CD7|* ]i&V ɻJ׳7lg}Щ=Y;Cnn*{1vfy6< .N:SbVJ(ܱJG!Q ځUK:E*V$HbEi]FGH) A$ ʨ9}6 ->!wF"Up*EkR6^h&ΆYw#ho͚׳7 !>v_nAhU-t{2agsާԼp!OجP򐣜fn&oPP Ad XQ*MRS .(#i=$ٻ6$ &㐍s`5:0))H#))QV-q=ՏꪨQ1,%! j@C_fg=͵_?>EK{" #gy6>nj7dǹjWx0ʎFBe?¯:6tG&bhj0E8VXLpE!~{LgSv sSˆ9_syP-ok& d4IT;{T{`?E1J fa?q?)#fG.M5t:ow\U?d;#J4M/}GGp7X qhFLsm7n f]Y0_a8V}do,=M5FK\}ݠj3TH8{ͽƷYɴ0W?_#"Fˋӳ$9E X7ہĈ^Юf}Tj< NՆNmߧ'Zh4ij`Jfdl^ Dc:q}<ޯA)ʾ~pҩd`m,{xQQmgsmrR S dN&&; ] 6uÒva/C5s'O,hmuw0ila`A<:n0Cn6?v1P2X <[Cn?ܞ=3a>n5~H>dDÞvv&5c*`A|FgL.Ϛ*[ Q|[e_.| 1`kfgcXfs]32u)]+p'j(SEօG"aaU1@2scWj酃 *Ű;ߑ>/!cNF[>-"ro0CJf4%f C"v<qq>SsKw}k4eg}ao[{j"qY!s~"qՀiKV.11|&Ɛ*ꁯLpl#6D({P5 ֖S~ b>$RNGiKp3"5K4nbi1g;Ja8R&fK 0q~ qN??;9O{΋5G|܌|?]xT6 FkLN'ĕgoqp\ 6r4U^v#!{2-߲*`ɛueZGo77Azf9q ӚIKnӢ&Jњdkk,QpȲB?ʫOȧjy5ϙdpxWT+rA8! 6ʸR"o696(Pʍ)92J#"( - Z )$1aM͆m:$ ie] Vb~.x$SGOgDLV-FPv [BjP-nv 7 HUjP-nv :LM/(%NGOIxe$L*| /AJ$|= /R /-3gp֢Aj#`Y0h &]ରYYvGd"tD9",x;fBb;#*$# )uQ.kmU1Q[BAHRVa IJVc&G-#chnDdce{! ̂HB9Uaa0%w llVpIx)TL'2 hGԩ)r0 >0)"8U*HTE6q6aCFl+"̈( "D| w*I̶JvEUłORbF\<;1uf<Ի &*'sS#7/5KքQ6` &.Kd)]}rߥw)]}rߥw)]}rųٳ,.d)6YMbd)6YMbګaΗ.sӏn|D2Y K&kyזl|{qJ5(3R,xEʎT_<:2jzTs3jc՚֎:]3 ?-hK\ƿ/K%^ a-87.IRwբ#K>{iMNޭ_Iw/NOݬsa2=wk5Iaj)RO7K94࠭&`. ^4F5X&"sH20" 4ӆPXיi',0ˬ9nF>8.GgiQ[(K0༭y"h=[@X,U$"TiMZ!EQIBY.=xg`q9ǂGJ >!M(_Tf<~fk XpZo#w}tGdF#maN`#BuЎȴI 2E Bȸ' "Zq N@:Ob6A(8Z$Q^R-Om z)gt5KޝlVճqjn}k v7ӕՋKeFHx,hPvSͭDq):}1T''@O/Ґ–b8v "-x<:GMpH(,Ӕ@[ yA,ȓuҺ?^(8ό/^T3ڀ ΘDHXOrҸ@LpMRX]>/=ʺjZG!CzQ۳;?@LRO t PRpk&5q%gQ'?s/)zhV(/0p%MyZbbCYE}<xza4b>6,A - !P)]HcTP[v|zqri߸'Zhx867⛷S=N>w[Y H0b̮w߯؍B o̒S?l 6 @omX$ČY5.QST2cV&jW欧y]'[4Oݪv~#T>y: )}Az1 .u,]kHu~D~ {mPMhO]}Mr!hE_HCѕkM{E;>۵\v9no:q~1VͪIidҷwnbF6~싂jů{cfvx16~Wطϋ侴|}9|} _H/q*%-hs=NJXpzH&ɚPb@xg͆" lp2o:b >hrik\4ihw*o. af n :[[4~ۼwƏ+?5kOjֶ\\NG|w?PKm*Upm&xoX}Ft|M4nƠiYi\¦ { 1$(« J1G% ̉Vǩ3>$&2-۴TxGDJNcRa̰4N+؜E \z-]; a5(zefG%{Mā#nDkIR}\\И &Sasޕ$RKwc<"/k{y݆1EE,odI%QȢȒJd,VFVFF|išVP}hAj !It2E`rAP(w"bxH1OݍQ貱 .%IsGkњz\\N D8ӁIN~s2be.1eCgm!畧"vU pp:`zƎ$ռn_xR>fP}538b\9M|(7 ?8ϋͫ׿?߾4װg|qUS^1#)[[ҁ_Hjqŷz0OBYKUbԘEݿ2; E˫Ԯ[Hȍ1Njh]x_ 6@T̪b#z2ҜǗ[g/zYJ׺s4zgñY2Q>ԧToPrc&OO.PNq!*iPO6܏⻘o^ek}+_8}qQ$~j>[^Jj7J1.WM(oȵOYd^A{]RM>wgkc˱wm?dhZ͔xVqjN,hym>hNM&5q=Z\8>EFDwŁ'Z;j@;B\rE\E^g'9ʥVCHy'BTukFq/12iA?E+W_(v^74ĕ j]Qm$"*l8Mn@[m=[s!F;ø,MhNWA#9p-k͹UkMADb"&!}hxNWRӮe9k˺""o|c/|M}}23ijJ0t2S9hOSzg( >bd@z3w$x(g6@R>G)"+P%Z3O=H$إJ{A]=e_Iu0T\*5c+m ;@X z')R^\iqr9t<FY;\Q!7vV`^SZZ&׺<+~|?usqQ˽|nCkx7%8KCT9zI<Q,$h_?XZknIޕlBs;;XYoWN`[,n,¥k,8}VL.iK].b@=hu$}~^W~Iw m| %(|NXRʈsVߑ~ǃ;@2'\$9yf.HŽ`h>RjВYw\Chc"NT.x뜡DB0*QYíN*Wewl5}j<W+Rж*_}6u|V\zqNZ@ߙW3{Ggj[ϴM{:iH}[IXexg6{eYO );no>ŝ.[ľ^sOZ;m=kwqƕx|ݹ}k<2`/kӺv٬kΧ?t.mE˧:iΤC>iZn͞m$Ȇmޚw,?oE)"wHdO8HʺzH)?c ."hai)<.gsSw;{ic}e#mrC=f{U@>v`PeH|YºcaNNFb=jh$Վ:w֗(}/lVPoww]fx@ʵ+ALjEFL&ʌdXGhk16uV^.0>94A.|ww-CZ2 ~n˩B+]@D &P3ϙ!O;Th9K ._!X:}dfˊ3X.Ќ͖!ek5^lubRwu&X*JL6ރ:SiL߃ N{ RXJр"L1E e&F8SА&*\>$Yg)2rrޖ",3LG JYh1*-X[g>;&Rz50*}0dDqQSScD ŕK 8 !~ < 5*R"Hf>#ڪ &4#Eƕ ޡ2`jtItۡnѱ"HYKH-IՖ}™X%e\:a4Jhe G+z7wjDi hD\3j)Nh ":F4ޅ@֞Gi]s .P6_k,-oEu#"ZY%XhبI1Hi ^ G{cdaw!PO z1Ɋb1xIY]!W@"GظI*yc߂amDp'ADJι+YRm >Q)Xfŧ:_y ,Xv}+cʶGZ{BZ/ޡ,L1B@rQt% !Sġg!d*> e!H|K V;*+XWUVCWH&WOP\›\tH\!ΈL.ʃGWfI^\=q)K߰`w.bc1c\>oe3D#^0kAıDŽ#U|߫9kyVzϮ6糞ٻS9m=āi!Ņy13vJqejwmum2 E#6ɽOrf `\IIesJtR"DbNGi-1 )'Kcټ:! FFdm:.<8kzi5\j1 փo,WLuǍkF:2ɸ2 v,S){S„ygrH[4p~ϊ_T٠&/h C\E޼p(Fgg}؂TScg)?7f>puh,xfL)A )R % 5bSRh]mm(ɭ+^W^{T׮3V狮3.[QFGυV()5kh@;tJT@A!T )کvA|s:i3[SV<ʸcrC)6{lnkhm=˽1Ee]7<}egXfa(o7={jhݤސQoߋ&^LPPM[>0C YPrv &N "NV2#@'D̽VT\AFrh1㲳σ2!d}ˊB_H1w*B?T'}b" Bq&92i<3#k;m"݈3V{cbfE,kA%=?q3GWBW5[gK1_4+rr$O4:(Ec,ϔ۵1 '&䄉8Ģ#l.k3uOmQ[5 A3"`}8&8zpˬR9ir[H$l̂YV1PJg0D3慰܂V+V즉.ϽG w sJ7{:geA&˷b Mw[yRz4킠-_8XWek/vߺڮr m7Z"7j4*3Aƽy̢ JD ?*YbP]V澶5Y_ Ž1:p.e*%1# ֱ^Ü*g1&W5AZ!qaN-b"+Ck=B"k:c% ^Vjgm\ ¬uCJ"  F|DI/5g@f Dzh} Ve<Qv!P3޿8-"c9DL:s2BZHwTՊa: ӷH}#dXK/aa1'!NhYQjwc쀣 pLpz{H:i!jgiΑOfn[?tGQ xnWLJ]/M* YJT<Ɉz]CD!%QE[vMY2@/zr*WE >S9{ FAάctb*-)3VIt,iLxkR4DK2Sjrhtaұ0J)M{J3?宪eL+^򬫚ӥ[3jZ"Qi ]sxkWsgs"5nqtu5U+&Dgk mI{,/iIIow8}xmK&;}[l ORJn> ͖rp%}SLkm%?}_nedicKwr4G ڵݲWUoݑW~ /qLQLM(8)^Ak jtkIvlyb+fZk/_¬$n3QdRp&,^l=l<./zT[vyhQHѵ\=r&s+-ſBnfъT=9 ǥ U̝n+ZTDZ]NzE ^<4s8]RRo9j88;<x&rN7x54itu`Nqb~g7 oofsqp 8i 5֗5V[TF  _0yOjSd&UD6dCd!蓓&y2I>s $fsG25R'!Ų`>+nF)8Yb k,61fnbtN!&=fBy+ͷ+G =Wht$hgs'Qg>9+Ur&YMwϙ4V`?9hD =5%P6('˿@~SEӮ"ɕS [1WdX2LU8f@˜"cR{(],[olC1p dJ!E˜x ^)[+jYYA>!8UW^_^NǬCbªl#CQ DbJ!g9#uH[W!a#$y)aGT@Ab,$@')rV3g-% 0xR%Ɂ%VYIϺMoFmPR{XONj[0ʅTfFkoKUQy%dIs9r:rs1lY˲G$Vバ(ΨSu)q6ɘYq.eAF){)`r9`1nd\ j#c5sv#c=RoX-8 +j£bҢҶma[f|g݋&'q:l0> +Gl YN$t.I8*r= lvD&|sU_dV#$@YtM)FQ l&mǜByT䈙EjF2g7b:HSAjq(jW^J9I ^D͢H,HԁɄi F Fkfml%D !WDKXE#Ph8F%IT T3g7~UTK !byDԎ\Z)Z}PC@#v:`l3 Ͼ'mАDCB"S;|(OCPOb[ӛ-.`G{3k |~]ovV٭ ztn#o}ݎE?^kR>Zv+7|^j퐉d4"4*hag1&em)k[vAwNyB{vʓڶX-lȠ%̚+qcY 3\@R&!lsad.*a LK%L$r̅@VVfΞYc淏 x~nħnZvWwSnsV<{6*:8r+_j +{r, N\B$; N3L KZ m2~ \V**;MF9JH&X'v>r|$@H J' f Ɠq̵gkq(yӫ1\iϽ-W?,(+B4cۤ>Ӗo1aGckX0]'0}$W-Ӝؘ8ty˱H/"7'o82W\GyD+~-[}\۴wA=;=FPICg8 E“4i5x &2/ݩ?Z/:̏Zo*其v=_gs8>}?*ҽJݴ8}` sAO@plό,՜ EEOZRPӌ]M<};^ q͢Xpmmk" 4Sz}sܺ4uu@EjsoN^_]~xԧBPocoztvj"G膓/o$չsZyE]}xI_} iYIXny>|0LVy"uYF d7RerhO3)h BfLkD8T Ym{qKj(P|y0t &,ՙb^d@i3$faD@\KvTאC@lpFGK 1x;Trwkx6]XDG<[/x\6d]Ёc-:R,2x&skpYX"ޗ+bxz: =6B8/A" B xQ8À%Բl1OC*Szm3#NjzN'81 Ip,cg6 E.K2u໨w]o uj%f5C!C,RT[ `tbyad%TKOt|+qý2ZDXCqF9G& OJ133DL!u:_|:_Rm6:W8scf^%sΙmBƘXiSu8ttsSxzzT 6ǣ;FZz4|3L2xYQ ;.[8>Y9O~.Iayr}Y7.[i6ˁ&4rh7v󣿎~ܰe@%]Vhɬԟmhz}?]gJ8B&WuW 0::d7ȇӤBJS!EJ!j$fwU:Fq?q˨Kb3e%NO<-60$p صzE fZER;4$] )~՗n.jcoX0|3 ]Qz*ͪ*Aauʤ6?KF~+^+p\!q`^8b"C$w8gARYa37Y&ѵMƣUV_lcJ?(W{ !Zɹ}\񥉨۟n꜎#zd䭨x=Jo\.p=-h Am&= bRxʨLn59m 6Aۋ(=tkmuw^;ݪUBD^SəOVeE[._mBRJ4 ֎:PpHVDWνtƥt,ӳ蚬πw*~cM^rct4.(Y hpS4hA f1C~[MhmKr{ q85A*O꣕5nV|Iq` IV^ S)#ueA,5ZL{7?nGquGq~V6I~U JK`֖ E<L*38Pncgl#aoJp3V!rcJ x \w;+r6ax^µTIOCw ,@=4 $Uכn u{κTz+O 'sKA3 2A8'#(2E$AYխI>H@b w)&`Q*CGwK I$>s,v03rM AigSԨcj,쐉h+D/ zFX'HqMj|a҃jBW!Yي%tH8f87 5fz| \ aha/*wW¨ПF{u$sjEp!$dN# "0l2$LIgB'n%o\C)g:Uhf(zhCqeFXL 8+ PU._IU9%grhx"M'a;Y( $T9lsցIp])zFw 97qྍQ+eFwNΝ6u6&XQ1G9 Ұĥ ,bLMGŶcd1C1 sPɆ@S`(h#$-@Z{$ܵ a/ݶYYMǨ<bڤ] UFRc2 (!/Yo *h4K^礄4`@fBnTH;hZ'9Qb2M_ PR`LHӣh M4\7ߚlh\wp<5ڮ]8? Oޟw4Gijf%e$B̛I(U83AQ<`ܘ?6fTNFA89=N_~7/~M_W߼>^W_ Ymwl {h/ؾ&YMS{%mV]m5]vyE7bnKOœH7Qd\-Lke?:1 (#&<]Oh*Fl8mzyHbڎӠ}ݰG|1{ҖOëJ0EtQbo? [z.Z:m`8nغ,/ڶQn"MOUI'oi"m霦3Fs_HBU>%>+]ō:h!'}p6ӵPao&0(a* ?5%_%:?;n\nkvh"4Vh҄6&ӴIۆ%0ugӨIF'aḜxIVlq+|U٨Qpd;%.jwxn凇%ͨYhf\?[Z-WxmOd߼z|Fk|; ,$,3I 6( s+A@ejYc.8#*f(wϟswRs F3H- ) , iCY9)Ř~o:+7Ďn]j{{O seX<я`yF?Iϯ?}}P=XI )++T>V\9los%NK4t/{ĥAҡ$^t.k:+B)_Ľs Bb ZfgQ~#Hae *dҙn8c9[$Ɛ)B|^hĭ(9qvft>v0Ix;  Xvm<{tTHk zT9 D2BD)nDUP>pPqX<"/JWnșaڂ1zPz}Hr"mE겍N |X:5ws5 - PE4o~=M~M9; =/]{q ʥÀ^W2σ>68O椞N/O-0}5V=0OtBվTJUMN]t@HRPU@$Y2}d:RԪuv.NΎzp 7Uwf5Mёb7 uǻ-ϑ6l|2oX> %548N+h8xL`ighWM}37/F+tх.7I aUg^ҋYɝtS/Hk jALo]D 2˺Gʄ˅^ʧf &/Mt^y+ ۭ@s%Ca}wR[?TE9~@gu,*()cH[cI1pH֎Bte.E|Kڬ*A=aa2w3r68'rAP. Lg|W0g4&KWPZ >zDj9lCG7?mTK *&V[yp Lg+"{xK˽&`9sƒO@b+HL1Ð8P gs WҤRҐ2Kh୷\Ef1Z̑읕΂~N]͟M.k% &^nN$^fmE[T!cjT_GRY@p{n=z|zVfuH7fFZҦ.vթ+@hsȠ!>a,f)]5;̂]bΕy#Ʊ]vިpg5/\&fo7gMLi͏Gy;{*n\%|T ^+6_ϊn듶:Զi`{Xܡ苒[/?ln5c2$QGcXUc1,Ԃw@ L ?@r)P<9R'fWw#WڧeWwiݨzށ]aϮvz`Q~<>|vqXN;ކNn*l4Z+*r|*`+7g)q1LshЛݙ+ E#z+ ;%YɘEɞ0!(rs4;N}b9@̣"!+1 Zyp8&=UOsX hZ{[Wr9f11k0+ YFZcVy1r/gX? O7l킫}& Ud@[!D=(=޻{Pܣ{Pܓ{PLagd9Do IsѰ ,nAM@7zM22ڛJ`ݕJx6>Ҥ7mۗb V +]}75i4.iR;jՕ@rWyQL߽̫^5/>~䠲]"~穠qdO Gc>YUc1,j擅ʅgd>8fqa،ϴYZ^AҢrIVRP)x{a~QIovTN<\ViM d (fGfH $h#! 1eǓ: Gm BX3'?cFeXb9)m>IDdH!ge8B/aWuٽ8B?%%RKR~_CR$%D$q$N3]_WS`%3u N0& 6$QbrxHғJ JujMgQkt MtVRcXmXqY+Qxrr8_E**Q9*.xx]FnTv* J@Zu %H#h49LA;z.E,RM'Rtv#c=]Vba:,|P,\I56.=#?ì&?mиᇋ<2bcĭIr$rY[0F64*EDm1}y"K Ą &h鵌l4CA$!e ~yZk:qb jWںCNݶ-gB2b_qR ݋0PZ"u^gJyN.Ru"6UT'bȤrI,:dkbbAC G(dզ[R?w-x,Xm2";D-HB!mu"kZ í;%86&!$9jDHakF7NJ n6\D#!I'͍H" q-3ݵGyG d묶KEb[5mtQ{V+)`h 0DAj%\G&Y$-.>.Xkw2 b/5W{\DFK]8'tEn.n}>+>p9*Abb72{h0x4)D޹AveZZVKh.eʼnE2*%*:J­^h|1˦n:9EZ TY,8HTZ+$jpC8FChj r]hFÏ|tr\gv B3jw=9YbJ) {|DfT$R nyG@N6nX6#׭)GNsm. 0Y:}>cqcT‘PjC6V T<8׾X:m#9Ŕan|"fB q0>YSv{D祧[x#s h(cGu6cQRY2i2YXcI2_cK(&P"{~`JcQ+7%98'lW?i[/҇*D*f)+ [At0 4Ml ʱF'EDHD^ˉȃ́W>1@ Z芦iDP` Ð8Qgs 4.GC _E>{\Ϣώ]Np֩KI]s";?i!os2HjPvu3m{Nolg+}nRˆZΝvgyyP/lo~u{Ś{EҪ?UvlMwr 'Ym\/䎟?a:n4dW背>T`RguJ79x :# :mfS3-r<&HB/ܧ=~]{5,2Rsfv'goڨ{]r1 IJAiL31eu !p ˕ Fs/\gqb8~}B^Ux .ԫO'4{LP n:,ZZS4h@ʒ:JWv`g8MUqh9Y<<ZE EJL]{RD1m[FfɆn+ I4^iJ8A+Lw$1;Hb}D9t_O6GFfझٌ2CɸH7ӖI0C<ƒdfqEmVX`FRX`F43V!rcJL@3ZADR^BVUΆjxǍ5%~I?Jcz6ڙ_/k!d*n+ \BGsR{0AdX QIvVbN]J(dI)Gw 5h,ѲXNg{W9&IQ߸0QhXA$ 6^) Xt4\(asE& Cf4yY鐘S41&yX$0Ģd:𲊜Cv q`j7F34itT^C) 䒈q:D(i!2=ZNK!(b4o:"ۍbb ́Od%>90Z?T>4>Д<59%grRFx":r:6Pc;h.߄h6Bqa4IP笘9TS(!,%]I);F= vz};p6<5:m<`rPV FG8GAp\E)(aHq.'?-WVF>$ נނp8Rq-ZgJB'%&§wI_!n~vlp/\Gs[<.HICɒMJm6f2p,uUuu}Z0 'CLBQ5]&N N B }~RuvtK=XdO]a^~۰z2zqqX֑QE_eڬKh"i&D/W*MŞh= goϚgVcvccq7/6耿8w?}wZn>ڥa(W`¼)K\yVwkpVvkc첱,1E-+emK6:|%^˛e|X jnXv]&ց_͖Xv87{ HO&/M 2^'0/ԻGy0 wbќ*{?V?lB,R*HwCt|F_V̱TpRVK0,>!>js|o3)h5]OW%z.e _v7Z#8d7!ݻb?koܺT ë6"uYK0|͒O५CImx5w7C雟fff;bш}ۆ;h 4v5[`y^Aa}{V~*ҾPzQ/~et{v)HRT@niNёWicxFBT6}_~Pd儑MUi.s &dy!X6방vpA9%\e| >*,u<-SUTFXB7WevW`q^־s<1`oN/{Lx'm%ԯIІ-.}Hài>KT_v늇IC.ζ~lJ^r9I4wj8Qz.]Eg7- ~]ZWl1ϓo1*>rp/4pc :yrR pkHHr߸8RJPE,?a%ߜ'{k~[Okϵ6L[LgruepSM3>ȷ0 `6sV"~~3ĮA`ME?eVrMWU2U2U:,If(L8*t!67gTd%Ti *ΤVEV錰aK1r^CqS VUtxrTEU,e;j b"v$t<$3ϳ"~Q}£}&b;I;|?$g<#SQB*,Q*C4 $'̆-z?]-յ8_}'WZޒe-1n7R *]DT;ࢧzOUkrUS)p2{,92ѼAguoۺ.ǬK>_֋b9*r#L)^e9Gnek"xV宔mrqIΞ;S]Skw 3 éfx%fXJ\ ʁdVJ*,RHsZTQIEeVΒg#eeVй!`F:TUL3SMU)Ykʜ?" eBW`ACDPFS+J*WqlLn i愥XB"SSE]SI83H}EX_E-VS6emr.J@JDE6j%h3!T 99BdvhS3wn:ïfyv!V)Q-8e*W9-e(y )Ei +s< Ny_TsXg*|썞K1qԪΕ9YG9 k32-*ӒX"Zdyڍ$]\d@jCP 1ddS#T1=KˆFV$ȬJmVV)ʪ`B:ʘշe6`bXЬkA.֑.q}=l{Hb\C`2U(!Vz/(U|PڐԎRD` J ]!Ze}+DiX ҕT֪ k ]!\+B+@)%#&HWJu@t GKr ]!Z!}+D٪ZjJteWçnzwTIM(T"Lj3&RWq:3//jsL"~Kwq?lY>TJyisӄO}& q`lfװW]8`] )ꉬ p󚬅$:ZC5;d-TKbm<A~m׽/0&FQ r՛ֶ!λoˤ_}x]ϕ ^[0m۪577?h2iL+y 4_yk >Ev^<l~sPְh7+r(Cô3!ZPE;pF'-8 N2 9c.e4~a=waІ3Ԋ_ ww$,/]ܭ*f2Pf{"IWtMw2]+ǵzUtR ҝk#]2} * ]Zˉt(=+A4tŨ4HWLhC+l ƺBVNW ~My 'ֶHG-aw=A ZzĬ߀Zu-=GDǷj鉨7>I8jX`L51>(=OiL5)/U,BBWB Bt(t5ARũV,@XW Pj·E: ])E 2 g}Z+}+@t5AҊpJWXr ]!\LVUmFD"]M 8}o/>gXީףF)A> H7FZNӈrm#:ThjyHG h['Z}QhUbbJ~{f|m]C;K?ڳ\[փX}jM=zf$BBWVQ Qjt QRDWXߛ` r ]!Z|+DI 7]BFCWW`+D+&HWX0B'{dU[WRG"]IK*B*gj;]!Jc#]MRʀ +F++q;JIZ鱄^R5cDq3=GDX`hPA41Nӈ!iz:4m!C-^-\ưy"υYDziMh[KWwwNuz:^CV(JbNV:Ja؃Db#`NZ ]lDXt_;v=?'DH>~ZL?v촘^pIm~hHeAW<վ]O2LDWX` j ]ZKt(t5Abf+li8tpy0thR*RG"]qn 8D` * ]ZJ+Dx ҕk\CWWP *;]!JߎtE:]zI:II#Ww%`Q$5oOW׵(ŎB6&??/tY{wn`PuMX gvHM;/7Y`,0;A^ǂeQ+Rs)M%Yvj XY:yD1Oő;{BW~ã1?~?Ŧ].ѝow/r:0x'v:RUwW7phM1i?~Ņ?60L)>6YOpzt1jW0S'^ZoGjl>}HɪO܉WoOe7?-,P3eb>!|^7o5L` q ˑPOǿؼs[rSoZ6W8VRφ5[k\]h[v}I*]un:[[1!h;F/{uo~;X伹^_{YW%qTIYB/Ng,i"\Λ;Y׏dJ.c4PjM!g\ȥ4S5i|{3Shy kkP $Z@oCv[E3C$ZJQ+ω6{Gb:0JHkfolfVHSzL-x~ <L{`UmɺM(Km 7 UY2dClt)CQ 3 Kh*gs#QG oT#FSHQ"_^WT$7u]]*6YtT:DyKP[ рOBr>}:oNB<A޴\s9 Sf0j$7.(]cp_^wF8'֤@{ur38:Aϼ#̷2ShJKgBrsUTeŵ1Y ЗI buRƗ ZR&b>iyZm$Bɨ|W!'c#Z$l%E6 VAfY0H!Q!s(Gh٥=yNڨ/ ޅ *)ёO,ܢ  TTci~mmVJʆY BRқC:_|"Hmbu2Bsk(|-S`1uV6`Ņ%̆4<ƺ2?K+5Rl{ XU LU. )ٰQbP6Pc.roPPCohAydSXb@?5:\vAjc4)@F&E0W  MGNsYbԨTߕzi[-P^f$߈&#L6B@F;=5(!Ȯd܁E7BnTzC{- ȸCLAA&c C='X !a@YPѮi\5<cn-$$XYu0wTL@TCJB*)ԙH͇*QMu8?v 6lC@WJo7XQ;S )ti7 U B!Ь=K*$~_H( :ͿBRAAuT.g-UuAT"F_|f(`3X KV#5:)%5s͐Ȳ(n i${jE}∽}Ai!:3aP R5٠db̠JUYa:iB0`9;ԙxa]/n~ۚ|uI [ŬΎ#D m3bka&a=Cw /M Td6Q t56b ڸLcuLC=GN`]@EE ]ʃJ I"92e(X LR|xLPBb2$kuk+w<o 1ttXTuaV% 9աk4=!V1dxaQMZ uY QHkC_o..q]D]eb:\yS0+tmi,#z ԥ$7mŬuԆHT*Qw}TPGQpu $$7(ڽ,=kpⶦ3*Z.fX;6@D]C:x XbU30be@0 9Q?mAOy~{Pqab2|5o9hE8ڡ*f`k(AdJ bdzl{!@S AO8IZ{m(=mGe(utqHC[E4bt~֘-jUeviL  kȎEЬ䭋*BER!6OSJ,;wR@ǵ1[8ru3Bb!>J+e`j?&zB?^ZܲKkÒ{HCjMuX1omߡ8}w>_˨՘Fegi %7 EPqT z˗ U%huכl뛾?Ͷ_'U.CuBk;k U\׼a?-c(*_WKCꝭkkfzu?cv^Bq']Oϧyo;nw={RB>KM^9ѰN-iw{|'ݗ+d[Chqw}z/}[/wRя4wZŵ?pf oXLA-pyZ@\'h i/+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕpU WҒ Wx1\Xҋ1\qIOpͣp+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕpu WH01.pRZ hf0\xlbz>+A WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\k Lt 2\9^'cb>,phҪ(h"U WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\jVՇ K^˭޻]ѷhwmU AJ%È\ZO^?9~1Q҂{bQ/VQy~sP] ]1\rK+FS+F},*zo7Sp:;W !fVSC Qs̠^)w왲R\?|6E BN/17vGDq_/^d#M'^r{vlY ]1\CK+FS+Fo^(zOt(+w/]=mP߆2>{v"z׸c?] oCW ٥Pe!]b0 +N..nЕx=ubF ]=CAk +6z9tpbAFwOrb1]=CD,. Od ]=CrQ9bŨ+F|1AgIW>'Z]y)>,Y̽+F7GgHW!zw'@XVi3^:c~զ4^dfsϺt;j/[ op@7: *VK*w~%^n}-W񶸼A2wW/i+BԔ7}w[%_o_E-;[ NWdZ2׏4vPqdm>=r2 WWWp8_zYg][s+}g x;U5[[5oS;/<Ė=l$˒#YLٲ+#w$~A?[$o,.Z$ 0l(4v|i5|z}YZuXt6TG>>I{ڪA{Of-U cL1QBZ?gg{e[ $(PFmZ2ٗ^(^>:Ĥ" jq"BۖM: j`✡uSL){ &V :Szݓru>s Zx ?FE^^i^գWmxЛG/Vlۄ~f1wv̷|H| ?(30OOf6pla6bNOE_E+1u/^?^^OŁ14:d EULTئkUR9kb /w/8!ujIx4e#&I{د])84򹼻nrQf|]$@->\wF>ߎKh/Ɵ_Ԛ ޾׸o[ccwNCNS.~oS;l.up W*=?4]SBP/Ҫ;Ի<kzILV4.C$v('WP^lEdb4 F?P sZwpsN_>N5e|,չBDn8)O1s=ZlLHd5 O6h1$5l¾s ~aqz@!3vk=]DO-1 }iU6y-sY+w1#.[݂֩Br<]+6hH +i mT1H(++!R%)jks h~"yoeC2VVfL' a *lo+ԵJDuoua:a]}?o `@ؔU7Luְʃy{zȃgߠ{<(Y02pјWѣmez }635q}x-XZsԤffpKןT˴AquCBhuAMEd"D磑 9ЌarYxX`g@٤^iT$6R)@R猧Βxm4caŊ\TRWQ}d C \.Fh:Y7q]lST>t ZʦJ^eci 9 7ޙd\Pٝl"ҠL |~+tcWEbHXs2$Dsg̾? #juZ8xV{[hӑ+<6f]mmpIm'lx`uJm 1J&D38.cZf7Ot;CIۢyϢ!C-`w&<*##CiU"`1>1.%@4od񦀢d$C/kH oFe sGOȱ: ރ& QApEIdd S"UpַDLXR l!w:8G?X17:iȱ;[+NNC^KMq#D/KWnetj9hO$ M^hΙS{]'-zxL0s28H[ BL[0<,=Z gm^ݶyuFYlc''!L }1eX3RMEVRR)~!*ؿ(Y(JWI}UJ5hRٵ`BgM?i?.gSH GYd-bt9&)znyIJPۮx&^w5>"jW-F,ßj$ d6McF\=l9~l#ńai4Z&Ίvӓ#:*AN9go7ף]ݷ+6sg|'>~jw#_KҐ^ ; P,U@"VP:"Zj @KB|%,LBjfS0`j̚RP+$_ȩHYHToMV7QK3cHp%Q{mC;닓0ߦ8n"]^^Mgoc[*1)srx8O%IeX7bc irn!.vytב|en'Cb**ч8=v<M;ڝ ^{@i|TҀdBaN{1 ˥UI]v}kbzWggmȮ];_ s` fhg_CEEt G񁔩ZȤ;MpN5x.~;<xoDc:EHYTt*(PRVJ.]Tr2cw3%ܸX0*V!6?)cWyS"c ް6#vgGrQ3ũX9uvӒcb(p5S.: PTITN)&!L<$1 >// N \*㎽Gߦr5#ڽsa3^ชy"a ~1IM^w\m|bfד/^M5"C\dj"E|AR2-4 (Nr̜Q1s qjj&Rb)(LZJʨÔ[:PNR"C7^=eH[堛R_{b!޿f[ m I⚃jz 32ciQfe+Y.Lr,rk(ֵDKHijȞSt%*bL*)h1k.*׀:ԒZ[t&QUc(o[qζ:j/]g/) s7 1￯n~Mgo}I?cXɢƈ0DXEfhUb((gWVQZ}h=II$۪bhkSZPT,)9RXB[RT3#e1bd-kÔ6f2S5 rg?]sGv]õyiđSZڄ]QN]^0Sr,*^ a-Wd#WSkF٤=umҧW1Jijo={F1|zT|GK` ]P$vϟJ4p\E=7;E $RƢdѦdΐ Zb > `;Dy!R%)j[!"yo,kI*YY(F*L,XO!o!(9PAH+r Id-(=anev.R 45OюPx/#cW L!XJ\zW|t*HB$KS 7:B[zHօ ө]N:Kc^Db0 jɴ6)FLHFŜ)xм۾NHiEDv'*1{K)y1Vߘ uCEh}'3>2Q[Zк?.뻇otfYu-Kh9Gdn-ﵞoɭߠ煖KoJ-9o]@$MtT0!Wf]ը5ӟ6j@i ̼cjϐ;I,6,|>qi+lEei*lGлUFnQFvoTQ1ˡK//[p, E ֕:KU |-˥J+&ur,w,]v,n8l8i5xG*j.ePn 25;͸x#ZMRR%hQw!KsHu(GUͣ:~ϯMR'Oۭw2UWLVVg_76ݗavŞ=^%%)3Fgߨ3!R$m,(QKB\h wyI8^AYN',ōRf-P|ʍp7Iy2< ptFd8A(Tcb ë<8Ũ)F)>pJ?'s}Ugz;2~YtLj;aޞ&^d)MU`М71@pzdVWͺS3:}nfi$5Ɂӌ Q )ΠMjKs>F[0 K" !aA(_hR3Ky;E=oKe_PB* uzc*q d0Ny4z$LOftv.[^ah[TY*(_;v4+&U,\ ,ch)l`LgL%fJidldl:;)dYm%bo^5ΠD@i4<~mGDC%^P:]Sr:'mdO;rwwY8PrN20hΒp32kuұ1k[`X6.dC= M;q[pVlR6UdLgPbZE#&rG%$ t`Q\N>lW i0<  Ndc h4M!).QPkVk(AawFº g4F]ICZM1,d3M\y? w]sɌr8>L 681"ɫSKNW0بe&hg TN;Z`Ɇ,AtchjRJV)4m MFl2hs͂o2ns+ƽ.>ٸ^ 3ī_/ķqja[9/AnĄbg"N)z0q.n ՘uT@EY)\78_M?iZU'pRV:2aVץ&@lߗ^إ/4|ݾ^cGXEt8MS9ۄr;OsZT.TO7tiյ>V+ͼ˗qf95 ZaϭfXWnc͢Eup/کV=ӨzVgWĢikkǖWPza*d`R͔˅r:%[cr9Y;S-g88iqReRMt* ɂ`'%O`hl,M86`8ж3S=o\ bkU%2 leaч `%ϊ `]F)|iyU'g6.oYX䪘jYe%=EwŎg.z!=5A eYr5)G4 O!G/Y8R1toZy#S3nx6€擞E,U]ϓ1] 7.eb_.sFj˜J]/sJ+`%y$W}+]+vp,JEqTA}v6TnRj jOK Xo+21TZ[E=mr> _X Gm`V,RJG(LDDF*ӯlϟ-@>awv ~iRJ^W\#*AUb¤6/n|E# f%ujJS<+Uة.`~Tqы▖X7oG+|ZEcO\z BW_\{$yT0FP9.6h}y_<(ȱ gԺ[eY"jmIZ#hkl*fItyet'ިjH]WՐJ:UjJl BjuB*•VkGpA\7~{֒w!tp J3,GBrBj uB*-9•\ P8طkXl"7 Q U,#bp/ "$?" n̿na59g(Jv聠R BmRLV5{ӆi]iR!L[Nۧ5$X+$WM4RkͮPIiU~;p6zu@׶#>n+rzW QkHڊJ%nTWwz av 6jo ȥ \!pB*\1A1{W@0czo {]!R:\!W\0Nl+u}WHycTgWBH&I9{WH.cl TgWRlBr8\]+Rvs+%A'W;l+BjH%ﴫWZ*NI#h \ER+w^B*uzpe YqK$ ∘# \+:iBI_͓T^ YT4?[V_8ؗb8H׊LĞ; Ǔ^W 7TB !('~~qy-?̼PhEO3 phYb2MF/81pPUUkOꝟoSoj60bY{R=}xK ޗ%V Nbie09,Nދc},,%iMGeU^INŽv!-%,JhI9pV\2b10!}v҅ngﰳܨ$}n;[;leNHGDE.E [8v$3@2.MK>&H'ba^ɖeږ;ΕUlVC!3cB`PǓs\5XFYFȂRvDpEk8ybWZ{p3xy/̎MoN.voWO2Xsp4s:i7Z4+qiW}k'\ hXաUR!\ Svb;""/xhX XW(}TblU1c+ּχb\CGWtk{4pEGU|\ Wt*;+ecbW`u_I/iڿ[?5wu'-. eeoLVCȍFƣs&'}~J_w*'?SmAGv6_o?-"qg ܯh.]M鵿.IFrқC'{н Z܄tkZn>W /oj\ȫ2(N'hd_>k\ R$OTl@"4GhN2=kb仒V0@lk~>]~jhZ . \U1q2hg~y`j+ AtJA 6 '%r0\rcЃ2rpAP*+[fbBH$es (}ztOecLڗr @]@cN#&(൑_YS>G_󍂡'Kw;,MB=Ö^SNᄅhDC1{3m-.wE9ЪOy/LOWUWb 6ԧ4i0YK^(ݚv-_<(3[//vNqUЏܨgRCf;%8sq@' `g()z"Z.N"~yҜZyΥ3 R32鸰1l,Klt*0B^f^N$MU̳R- a"p¢1P!FI1-Z;Y9Ϊcqf˱NF,*h,CI3 3gxaٞQ14ZzLp+!2.j aYx59=GI%rY_֮9+X ( )~cD%\ժFlN'w>q>ξ;qʵ=B_4>({0̾st"ATedr%,τ* Աj.YB tXZ';S]]>6ڙM1`׽O1VbH|bomr|+S.4GLJl4R/MDЌdh$O^2^DŐNC袭ވv5AuPCO%Ⱦ+^Nș,0XDSAI[үYD$XwDouQf0;]ڸ\Yie= otڽ6yu502`-AH1{e@M2ET&bx,&6.&Juq?ǵG0]J^Esft]7 [Qw4鳟jmM?o<ߌb4nAI]x)muN,k#=SڽR"5irUMe8W>~6nAc:]?]^\)˕ym}:Vnl;aFP2{muQ6ד~ fpL&H͙0::%xޢ}Y]CBZAFN#tQ9CJ,2BFZS!@:3K<aV81R(J  \(E}?a"ՍD2WNfPfJ'dz< !Y.p[|``s_|۷ԣ=Soy|.8i4q #TFj{ԁpq,:yɾh+p6ОN+,7)#/OQ&l2)dgp%p`xb-]efmYkR1kʘ(%(.$rTN*τRƔ$"8yQTֻl=7FT9s(y<7/ ',ozuhsYPns]Wtmd,[(&i0MnmDߘٿE iP824ٹ@̍Wv]KIH{R$!aHĸ*XYReZ9Y!I:{CK4i&Օ.{%ɻ<]*ȉX Bu!I5kGa2מetˇD9v]\.Z}>yJ釅cǟKPmcÎ֠a{7Z'1_}+='m4eiWc$`'A&N!)qMYW}PDRՔh XUw;'S?/ZnKR_/qJ̞|HgAtU۔c̗ER(oJ2Rx,кתbP/fHdyf^=0t J?ВYwl4EF>UYqo=]7\xsr#hlΨ*6mν#U%uezs=l njVb[wiy|eeZn>nu}ͦ9os淊_BϦ{:nohOíYoug2% M2lםZ4u!R% ֕ΰ:A @u5/3N=-䆓p*O`!85؃ei*17Uq^)ځ |,*н>\+tH&SzNu!frFGN<\k4Os&XC D)K&DƄAd !SP|c0<a(3rCB7Рc(X^ !UQ5ɽ6z"$<2)O''?[TI*q̔z4!(JXTPx@i DH0 3=w7 hg?0tP~[H^ꂈgn4~g9\KP'H7,y !.;n<|m?e=gf`NTZ=UbçmU㷫'6G[Cg$%1HNJ?F1:0!88"-Qju\ӬP$^?=ۢ{&V׻O[ z &:-dHЛ,lp+g񺎴^1#]s!- ^9l ]ӥr.պ.=` sΥL|)bQIq}G`{8:$3V>[E LBy,h6,E=Шq\ljyؘ243HM$K~fw"~j G4ڴ>&MPVS15m>;ݚ, DV Yʣˍ+xwNhBG8vJɠ$G@Ai$w9Y9)صLL=7rӪO8FQu8[VŲxeѥVG^:T)J FYb}>t<(]rxQiɎi=PIT (),tf s6+4TI=J,Q$u M[H2<9i8atD ,h10u單_Të)r<?FނE?C28Ӏ0_M<62bChVbTNK(!c찓D^9`QhPFNqH9hԗ1L@ k;mpV[8(Cd`.(CS0$t]3p 1+)K-C0H!a^ *v3:G'Qܤp_ 8 /ʖPAENSsl ".Vd2"34p/;Uu0ǑAq#N\py룍.\O!-SVL?8pv0eɆ 34% \$b5$5, R'M \:=ߛ)[ʑ}h/" m1@dIʁ Q(A z$"1O[DX7ϒN5hM=n_$J8ӁINK_ dO(#?ϋqV8WO*KU>U?mb8^2r7H#I5WUh*~h1͎teww]:SS :ow;뢮TwůugQ8)ޞ~7o_/oN_}2}^} _pfAh~~Y;CҮڡ -uیql>n%U⇫?_7i}zV*<{2lӢ.K5ӎBW:@]t0-|JPA^XpZHQmlԬSѤ0FGd/oV<9|U2"QKg9ߪqio"R|??/coMt҃fwmR_z 7ѿ/PNgI4ȧEU!pynwJAſ:]dm|5_Gw(ɸi׋˱j k9JqrzUi P˻sMFj8ڹţz&f A2kBd\ɟp}A~mKe?:r-Ewi(Cq)Yˉ/^ԲՋ6~Q/2}h+=Ц-.j2(^}sqUҡ|Gj2r2tyR4ݷQZދ2l9\/!;RE9\ }L,RpJ@^kڞaB=FckU wWn;uZ\fըB8ߺztvb MO1b7 LJ:-Aj}xis{79 'Aa*b=1p<5R+QQ,\9|v{h|=ZcΰxC~XjL5>h[d$Vp=8>?P vzQ3i\ Ta"Tސ칐RMq RkLvKTpf5L~}w2[?aY2; }L2%A@eD(kQʫq:!1 NAjڭUznl $l4IPDBO5Zks 1% НsRrB&`u|R㻀au+}9YVc] w^|&Sh&ެvkpck.ڶYjpB$/g<@ Q,-#ҁ3eb̩Knt xNS0(Q4:z>XPR]3a ܙ ׊NH[#n=(`RO$!TPezNSouŏZR(QC)ELjES.Lɬ`jDzMtlk -c'r2Yx~#ʏ[b`_j;OxkcN.Z;(sу1,\*Ц$xHYӓS PI4`-喌K]Mם}Ε!7-Lj3t^\ ܟe_]orW!@lK],!)1H-;ATC<4$i"{z﫮2v_?[SukG4#9AFS=S:8OEURvd$PNxƮ[6бm$[ͥͷY#K 6nŶ1VR& 1,WPãjnDh:vF6jMrgTR#7.xy a)9I`}:׹A\Ν%دs'))s?s $0'\JJu*ItXrNH]%%:uU'@Zı$%*z]?NJqd6z u;uЧR5S'J*r*IyUfC3JNUXlhF4-BYؔ|l~%1M.#9CVL8A7atNu{sI՜7XLQ^=WrҤn~.?^ǯ~5oMd>m҂\/(!#q媔ǔBQ" Rp*vw瑖""pb1tv"41"礱 삥EGCVCe97 #$Ă #52(BCwZ+n'PY:Dpps DIϓ$=-EF&Z@u{~b>iS"d<;!*b@b K EK)`hE% kYk5qX ǴR3_) U^;-TR8KAB*`3K [LRPh S=z`؄P`epBS3. A 8`kLU )d4&j`\&ra3t\R'FGqla'# &Pv#I!9ArAD9 sFqQp)gb=먤 ]gJ}6>-47,$1a/𬭀_2!Ǔ!޺<̍&!ړ'2#7 ɏ(k $ 1]#GEdR 7$PJ2S$*4ΠўAcM;8ocufXj4_`يa/8N%,:w{Ӷٶ}HwDQĜRafXTqL(``!AyZ [{]c o…#ik bY4uuM@z\И Sasuig;6Rۇr,h )GHwAǍOqD{W0аO3(h.]p]`.@^?{TG1vXKA~j;V iJ_* ;*܈%}]xY(ͱ#sODFph{&:H"kKvGP(c\o/j:\d;tWWTs MѐC;j4܂uS> \9xvЗc{sp_cN:~xf_􎘃kKd$o$i $ݳO'Ԝ]66nz"H%9O}Pi@M:C9-&K7Y肀sɓ͹`0NvG=U?zDhK}3(ST~mg<]egI%U8`-%ur#1cb.n׏rRn&x9Ȱ,A[R.k%!RHD#i֚8@C %c۸v|Tk7~.P{1Y*´.A2 ALC*/tO5^y4t*!.JIR eg#LuKˣc12rPrI'hIHSRlj*xύrbJcpWR`B<U8Ŏ SZ D佖豉hj4BS5qG{a\smo(edTQ[] *)j'aGRԙ@p>Pu 7z0uBBnm:n4ܜL)j7IͶ/So;cbș:;fr2ú< Hs {uJSNPK\z| `egUaR/1ײ)V2""&ZH0<)c"-[&ZebdˇŦtL&kBU` q8mFϽ4v_*vb{_Yq꛾m4T#V{>(E!C:hGXdZ 2hMAgi.WӸS bx9 8#`QQhlTlAcI c6cU)<`):=O;g0\NU=4!{-b8v  xuPYHI yyZ< wս] ZRO9N~[g gL"p$Fq,MOrҸ!15uJau[[Gh}XŚ5zH@Q̮gvi.dKx %pUh ;ոJpi]D4uD2GBg9Wy-qV Іrtwj|{5KZ5UJ^n zE0hJl1JWǫōR?ɯuy4n2}/֋d9̛H"V(ě)[a0Ix> *{Ư8f3.o/r{1o&kmFE? =2/ },vdˑd#beȔݲ6U<,:-(? [!NƋg%vq0POju3V4;:pڝ%ߊf;pV}*'Ǔ"4?k:8t2]%˻ţԶt~tŏ7ޟ`*R:oU@WV7ri#&AO64O!`͋eG_%|%4m[r[\ͽ:Z&u75d} fߏ^LHq 74ڪ^&-r4i657<_T+!2_g|؟{5v8GZWo4xe-ŝۋvΰ,dޯ!X=r}Ňoh=-yy9ۄiSn41~^J|=0?]*ݟ_ԍţx9+OGUejO<#FO?i{St[jBv)8^ :KIeqmu_Hi%!۟}{{קJk$w?ޝ!+Ƀor&ulz~(Bk@'3NRӥ]x@B_]Hf?,5{=FY0}jC3i1v>%26{̉ߟلnqb=ٕ75NWb6`݉{6v+m9s;׍9q]/I̩N-7Sw~}mQcJ{uʾ{r+ޗ'G<7^)kbY+B!Wi%tDq%+EG΃묠{*j__h| 9_o {LR)Hn#e ,:t2hhSB ^遘O^7vɹFMΧ {XQ*_K& /$!SzoEg(ͧ"mAWͲpaC #(לv IrSd4kP" `ҠrJlX[-B%4Rahm2xǓ .Bbhs&4eׅ[#s*< jl޻yLzMӵ>-~Wf7Y7Nl1P怒eu@`S,ԡ/ZbJgboXT>xc/Cc?O `cjB %}NˆK9t?[Q :$-/&3T*DDU-r+`9l<0șCάG쾔mc "g' ̨4O1U,Kj>g1(1GarYf| \8,ڑD]]&q^3T toԇy0BSmrUpnVWJe~! L=w(4FrҍmꞘ |)jzǕFwHA#8r {cUy+o+]lŞ|e5qt6 sOL@tBkbNÓV:a })*cpmя$ޝ{ރO!5fZiځ¼qJ/wRb b\,~*Y2A7**l@&hGh. 1›>3F&IB10"6WSyWڣrfw_ZKǠe ,:&F4$^kyWL/gXS.@rWʠx6_gRCfzwJp$2Б9@8s؈ϭ=UɃ?;i|~'b)'&ƲѢS-eJ6]ϲ/7aaN9q &m GQQdelS2V$NVjlg_}yn@G-K[~ Z0˒3 Adp4h2(Wwz8 k *R)0MV0 IH"^'0)PdIOv~kDZ(QaALqBVͅA9K#yZZ'Q%pJZQe}& SO#mi`Yh&\Ќd$ Փ$ s8OE/l%htIz/"cL32#Rj6M](F8\dIe+SL^ ĈR *C1+;j5s؊r2kQrڑ_m&zGl1mqR p \L-ndEIZ'BX͜-&zՁ霫OszɮqVvC\| h ZȒq2q-@+QFm41d"g>bopbޱkeʼnE2UJ*zJ*DR X6*[6FC;b6ۛmp{BkcltyΚsRqW1S( ץ.(.Dr A`gBp0SrT`H 8yoSw9$Eas!jɩ2k(y܃Yg"o>vC&s%+f2R1 ZTgm(p{2/fw1 $ᒐah]c37\F[c%!_߹K'I`ޕ.'F.RZY3TVOH]ÓQWEܽ>{"C+" 3LUWO\]!m葽dJZ|tuu+G&jxuu;) #]]MVuuZ Ο* ֕"q%WOE]i=tuUԬ種 OφʰCd<5;wqe+"|jQE8JTYdN6F\xyy"UY茪[ȍJ+&-M˘ggit/{s}H]u> ?Epe;^Daj2C$n w+eRFhpkxTA oHa9DO4P.'ߴ1O㯃PF7??wW߽X\-iaOON{Gn|pxw>wTӟ0qoX[ͫu6>yi˄݋+]mewiSvڶhQ}c2LN4^kJKM%F߇5P5P5P5qN`hst)Ǡ&Ztx&*i F{LIm^N:qTR~ס+[.DS-4r!yɌjHr@eQ!*CF;yAlH!x;Y0$7MKؐ R2 ;پA\Av 7B+zE5 rW.ʅ^+0}H]kʅ^+zB\ rW.ʅ^+?PB\ rW.ʅ^+zBZrW.ʅ^ЫF\ rW.+zB\ rW.ʅd O( >|7Pn맒DXz( yn k_OzbmD|#|gi"Q-wS8 MqBiG]ȥpL3F'2UpQc,%u &HLYpcrT`8 Lf#);vt&ΎQ#BC[͗IGWvlgb.&2+Y ,<"ҥ)BPiy7@N6hY6#NWp MΪd9W:$gOgpr`]v&+Z;;Xǣp7-{ysSLbqEu=#Lrsd||z llmhd014+JO=y(iB 9 wQLȃ|Qz̛,e&CiT-فܔ;)dq"`'A( )'$\JOftcp0\)ہce_-|uf[Y5c0cv3v݀[5bH*X)^&{3ht'!hR\7R `Q&@"X0fCyLyꀌ<Ցcj%Բ$)E#(u,<뇯c2uC"I?PO# xx1 igt1|h-  :\H JT3sRyAJ, #8'"JZ2@D-ӵ” ܥD6k9H ;C`IL*&@*Q[eh rH+JaB VIK&BQe{LENq #aIPM>$ 9$7u 蒉d:;-І1r -޸K;vb4F! .$Dn$M &M`*{U.Zx;56JtcgЍ~MVqnqUw#QpmG?.vc)W ]Sr:'6#9xpR;cum0!ɲȅR' L̜s֡$8g` >աѝCcMݷo)u0,뼘 "Μp*scd%.GPhZtmm[dP1&5&$pd6VbFxxrXV :@k{qi*H~2 ڤpJxd!Y.JdABy:gWߡ0O9m隌61PUNhda 5f&v p-Bv!)aVqϱlH S"($Rz&rx$'S/ s`{羚oӤ`vp8j=.|^ WIλF6d^7.y |(> {BLo{šs_pnJR{O+_B4z׷'o^˛۟No sOM@r{ pgp|ݛ漫y]]zwiWw ^]>â.퇅k{ O)ZeAlԍ{ӨTn{qi·fU\+䕧19O/5q_.RCԛqtOFy_-2+ˠG 0YhSܓ}v*1}]kDl׋apg/zZy?u#ŏӃ_ѫoNS,?@>-[~X ѫouZz+ucǭV ̲?>[-5L%:*c52-tUDu]s\SGSd:9N3o)[A4.ϤGwomǯd|MIS+kXq:}M}vѧBO+]V;. g%QqkM/T@襁VCЪo"Z7qM7Բb; &YWO+R'47.S76Vtn^<aB 6YqYZBW^6. 'P$g6%\6[YMzNk%znccY))C/(4~РnCa;a27Hɰɸ+`҂CdqM]r)`iݳ5t)(FL  ?*"*k: V! ];'Dlb,|/ 3Q C6*Z5Y)5sW쐮͙*{Xu;m&~ߜAOt+Ӯ͆pXG4R?$dRq>EZھHTMm SKR~ ž0vB̼J*Y25Y )XlFcОb`:5~wmPb{ib.t8́ͮyEUpKt>:P 4xmܚw }xq4; !>J$5Ի ݬi䀠l9C'K{gqRs1wcд^&`^7>H&@H2&͈?Qv`^L: 4HlѱF%% e%NF,g ($M[LyEwv}e}s5+>T; +ȭbGsn:σ|Do׎ѥ u ̺FgTd7B&HI?R!#[2ÆCp<`$ P`-X@ Pƣ8*!5C 4z&@Q))LPʔ7>&EVHPU$$gD&2q.3۱g3qvL hk)Wpo¬d^n~_lRvv{~/SW8`D6:ˬEό!62KkcYQN֊c.x)#A4B˳i\CU $eNL\O, %R2">2-Iws&ܭBNEF -N к\{2 ;ICgp֯*`Mo6ʐ~%Nw(O]_[;w嬁-|ПgZ{8_ l둻I"7^~_csM IV)AF8p"Q3Uʔq: O?3ģ+i J[aJEKJ>I 8#PA+Ƥ3'=Cj[맫 W?|ܟZ{50x{68gQr¨:' 8L>ϰq}0_<Z='x8Z~~8sUP~],TZ(-W_(+$aDZf<ӤJ4El83RdžCoàj Qˀ%:L|]E@<19;KÚb†r_\Eŕn8:+lH)Y&X;e HWPX/;آv㳟E8wυ2Ňr/^FR,>_}k{Z6 [z0ׇ.[?NKfZ0KQ_ȯo#'oG)m5?O&g!_C*[{eB42frɆΞ-v.m!`f?TrGbCK,;NάM$ṁ=4on9rVx.LnwD#Z 4NOfzm*TUTjH~ަY3ji# nrSB#R2(iRɨ# NJ/h JX8-^@_/;UЭ5سƵ6 =Riva*|R@5OUbRl2Kǻ.>\u(8<|>-lUZXZFʫ2MFN8SVL09s*)ҞIɂoV)26JG+skhgOtU5WA*ïT/c2%4ؼo;:{}Yuؖ֗$vIS3:B;"}l ԳfmKv߳hJ֙:x'0BB;?( N-Ř Kc+U6F}7+DɺVlpB"pxj'zf'> #J:f]t=]iIs>0 ]%'C9;ki{:b#2+Dd6tl R*NW eעzzD;Pw :Jhq'@TOWGHW\p6OQAM8 `A%dzLF(~Z|2.%޾(y3qh΢7jhڅC޾+O?nst6`˅'ҦQk|OGa}7NJ ōyŭghp< -K^G^@VT p|__R⯅Ғͩ5խ(*J"&IÈ׊< }QqKb7IZ^&4аw;/QP9qAkv8sY׶vMRF^xгξ-R>>l5Lu'gZrPթvB2jGh i3}y.t%RA'Jz:BZ 0|&\ ЕLsLz: 6Y"@hq~;JijS,ڼ,>GzIqfCӰx⯷o6B> eQ:GzjN?8iVrw,!~z5g'~BJ˄8-ޢ&t9*ԁ^JjK8Z;/͆. $8ZR4w byzq2ო'Oͯr1q_pbh%QWN!n3V JN:A-0pB:8eSWH 6kn㸲ŗĕ XJ?!٭b  rm?`KCX&z3o{nOeq1rѪi|HjN c'TgK4ipg\hHGM\ DA@1WAۚbI$ ՒH5BysuZ N%c0$R%&u+2$뒲hѵ^顁ܟ֌%Xv),NPR: JU$BK#c>Lf cSJD34f3NYVIʔE8?{=-H=׷2dYZH[ІiNnF4**RPaf!Mf"\"U+V#IP2R6LHpe &0'X{ RHU(NP  * _tk58H&`^ϫu J+h#1wl!ՠPwSA02Pƪ֗|k<(P, Lh44QD&+eB>7Pt,,x:M;‡ q b ʮLJ.*)ԙ!Oź*QL-!3Ф~ζ+@TT|FLF(J892r$XT5GlߥD"=doH_(d[ ()%Ni3VW6!;'$uh $1bc?V%d0OA:U KuZ"5<) %VsȲZـHzҰ WmFՊXkuڄBW;A#kW$' +K;L')Q_0{0,b0`|t}YsXUMے11-u9/m`-$ >:FjAuPiLJ7#K_^J.G[K*d`1uLA-ENp`CEE ʃZR I"92ʘe(&Պ.L%ȚzOE# }IEHV/VH!3XhGW}`QչY,TGW>^ՀXż3rmE5Yr&e5D v "} d~Z.ߛwٚo1 T.@e!B|u)]LLAjy1k"aB-r!P-fH Hޘ jWΚ%Lň Q y.!z b@ l+ jw ,u惙U3`0bduCJ<+ G I/@Gd⢯k7 F!tf% I,Bi9DHTy Z{(o,*"Fp(yXT*üO(P>&qH4bl,'r(Zsl<}t֞Ew֮ipI+ j@ec޲6R*4[$XH@$0Lը[ݴײg *&!Sj1=6Mum+Ϸ |q9i立z˹W&S`A.nn`3 58)]Q׎"'4Q;{+5f՟Fj98)E.Ę43@9[Ffz4L}YeFhQ"}_AI 9)1lz(CG3{_N:)r]gd*"2 % )􊠔5z6m6`?f߀yE!XIBiP'W Wd!G Ϻ]F F-)B^e 5JuT/FB,&z΂ U @ K6"\D:ƀTm:c&AJ?4fzPJQZ^ڟT'#)ɐ+*-LB欭O֢~]Y%kP|P}5oj+F{A DO0+gQI&Ezʨs!1>?MД nF5>u_ͬBS{p*q˾ %..IR[1xx@A4DĔ5*5K}1>옄j~ƗP$pԅKϝq%G@B׸r^GtE#Br7z)XoL0H5宷z_]^ܴZm[ǽaA5W=!35V(N_|9*;5@J EP2ʸ짟9 tU%_]l%vb+hCueCMV&?R/6Oo/~ѻ_gX7ҫr}TT< /5o+\|.ź_i\^m:o?blΈpѸus|}~ m>ǺYsqlf1a;/Q~PtgJwU8a8&'#:«=#? ~N b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v=_'6cr̲x@{DN 埼(Y:(@ݰ@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; |@B2[8L_ <'ˬO ޲9:Bd'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v:EzLN gV\GZ(-9:CtD4 GCSBm0O 4 i:u4tZ骣D3];]сCORh ^>AUy?]= ahma(+tELW;}wz8&`鎇W q,t'OWLWϐV#ع7c+uJ?u(9Ime8"ꀏ:ܝ:%:]u; LWχO6]u ]u>uޙdЕUcRWl?sU:%:]uKϑz#n!\:buO:J#]9B ^dG?suу@ ؞v_fuuuju3Gߜ\̓5N*iɡ7h%GQH.oNyq7o͸y}>v'9}Zo.ծgUM|֧X+U·1jw٨'3?8J_ uaiղP}C´JJjedR..RU?MOϛD20|W!Q7 vSՋ%urQ <;#j°Gȝv'DH:va.Eu@ ^ozu]X^gʎa1nTZm/$犾9zZvr˨?ӯR/IBs NþlުW=͎Y|we=gru),=&φ6by,Vm=ʠv]УͩZC}Xr+Y3"gdN yG\yTp:k.}᷽2w,^e΅r;f$|VOrZ9=e=ȧR&}jQO uG [!ݢmi=b޲_1&j3xgC6tǭ{Çc  HYLn_{O]aֆ﬍2з[Z/W}YM7Ai*XcUۿDc_ >X{ A :Ч55㐖VCV=y<EG<VEVN,1 r O% ,F(YX.ڕ ]v;,gH BY 0LBBWO%KXzBrwRrb.^y(hF]^Ӿ( &rA?]6iUuuh_&YgyLz~zlT9 (10R]dw}Gy鿀[`_F0Sm:Z% vWzpUKAe~,&u/g=llx]2'*J|h`Cby1_YlKRf3~M݃+^眱")o.*Xaަ驈Aqb^.k) Yg,>1,Pa.T}'P@dB?CfnόN.U Ôq0E$t .Ill d<*kޅMzZ*JbrmIQN3{|x+]]XTg4V\j7S8&(X\Ӗ;\+3S@ ,OZVN~\;=?` K_c7w(nF76X, &iWui 7|=B o~{U[N"~zVx gS> :xkAx TL.㝫Ϯ/=@4d*!.JI"p)a^SGKU/QxWq)Yq\Fwu`3ٽe"gjRG`0m GK,uv:;dK9/әzlRTR )p4r~* iMDul'L@^E"c;\)N95JB-ra1ЂI%T ˆ^FcD2 DD꫻EAH0<)c"{&:CgGlPmJ(Zci7iP=mZm46 1zl4FKv~o^lGdF#m}rQ`#BuЎȴ 2hMAg(D"$Zn( 4 QHH 4 E 2<#@v`2ح%T8 yS89\SǨV39֏LYU5<х␔mH=H)ڲ!>#ّԅO 4Z/\{EAQIt)sY!TpsŘגg0 mH8sՉe*)WV~u {ggz{E,ΙZ<7c7;+a^]:P__ûwafxvko0:!Xg_f JhJFQ}yR0x]~ D]뷟CM7wU ɤ/bh0)r0aBy& D0YdXiMO)e۟@ty{U/ ?N M#|^fAUW}%AFgT!jdh;i>[W8sS ڷgٛR>Jî6>D3n;)ҴW_ 2K{7^TFeUtLA}USxhQ;bcno?*nڰ^d7 {e'ꦟ;3eH3zm ä Yo\i9WjۋJu/Y 83E1h| M%h5ͷ5% ~Sqs.Gʫق%c_fNzqb$m]fu԰yx:4-|EC[2sIid?}[(g]`NUh>gRhT;V-dn5gyv|#,Z5=%nIݑϗ"'#R~K7k#Fi0~ 9Tv\)㔆abJPo`ݒ"غ㌘'iKQ]# V[M'FiMLN=;dMA{P54P ;f L;:g .!GϾfڝ3O7oqm)voq4֡^xΨg(e1Ky~='[y^~lƵJjhn)9T3Mpg\?e>2/YP0Q_{i)`tT;&\e@7A/%shө !$ #52(B;NxMwvvU=w3R{izsIH}U=5?#߾ޭ.6JwJ]1!1e iPc &,Ȣrrb hE<4F3) a %M OY* ʨU1kQRk R5y3tvL]VAS) %N !"+/[.FZL4ZISNuے@ل(]!-Vd8 䰊`c,P!r]; סw)f]?*޳h"q'Z L*Z2{T(h$ K#HCӊC;1  V#h$w<b8%!i!Kvdniem8X<*rFH)`R =Ad1؃JRnPUaE2|e&rRbP pPV t({KYw 9XQYge WUXqq]պ?#@r %)>٠ cV }^7aOW߿W_\a^}⇗@<{QqrtҖ;svVgV{[&zXt&*1sv6npY[ Wˠ@۞Z>a=mL _΢uѺZӦ/r`%h3`]ɕDYϩӈr-8 FKRt\zHV+!+<2nr)ˣ * $ƩDHX뢷\Q"cP''B`M`=E9<g+.UcA?!0EA?9W׵94<0smCx2I)cY#"n {AWCGbx2 Se^t',%DyR7s6hƤezOH06Evsk,?n23k߆]C.3;:;~w(reM8ZBBĄ׳$eEPF;CbO0ؙ$$Z=zecWN/׮DEv(ljl~+|fc/<Ÿ.w??"+=6b9iMɕז"e,wE^✼acݭ7:|~|MpX=+z ;?Z;#:%I ?:k݅JF=åV{tɥtTOQ*;JeP3J%б8+tjn\?5\FĚcL;p*[ z(v F/BL9TktUS Y%rQF ^}Zj2W(>;S̳^ɳ֎7G֯I޳6GY?}*֘^>wsVY+LDъI/+Ճ.k2y,.|p}#rb $?`GB((d2:h'MB&e# e 9j+RQ`6E)T,q M2,bHRKAYQRBUs浛9V^^Ɛ?(jׯO[z֯գON1_3 AC,ɒ.9s 6>.V4Z&^3P'gʪ })vGih=O-z j|p,ڝ/3ڝ/a;_f%; ;ݹ'dھ|;jn·w}C?`-4p[ XQ`*{Ysϵ>J'b"M*7r΂]rOdb2m!KOo5*Ag?}n=qA Lb oCt Md&Q>=H2<&1t ޫ}E\Vwwad!m5ȝZ dH6 h Q'Hj4xԾ1yp#jVg +ốm&D5g‚uਝr5[i($*kJ@W?KN>U)m5cW8(uύ9Q[X9󪆡cE.cGXӊ>+&Ӻ4.5 ϶(LBP*Ѧ|c$>rD22>\pf~33xgP{"6^뜩Z8eSQ>Ӫe:ς80 \X@.E "xTH mӕKqLQ9γ4˞hzDŪb)9;|FxvƑ *p)L29gH.*1l X`P)0.PRV7f?i5dmPmֶJZ*j( +i0gAi8F%mL`o±xLmV8^1BМQř:z ]+s i.`l>f hsL)` /ƕLv$v3gX׸rx+/?Ws֊][5 R q(QO`*aPkt寮_$TF}`E<` f@ZG<عaa$^֑JY_*,AP v]먑`}׏WQ0{y}0;-ު݁ooQUgIlH*d|f[yB)32!R XNt!JV󾢦mtb! 1: 2͜(4&s֞ғ'>,dd~}ƵM=:r?MN2U|p=Ξځ<5ɹ]i_7|A8MWOm!UC(f,iJg:*yD5:ew=zuT`HNLʚCbd*إRƒY{ 1NԕRB)d+m6Q̊kĢ*jS5oMgb͜ ٟ1u^SI: ;rx)8.jAb,&J9hUMZ"W#<(|M2OV6(IQ&v3mN0%u`%Z*9_@mlz@}寯려A'*cj]Y^oj֋^^SqqjL5q5;5ؔLuZ|VOBc9՜6" =%rlC\+W2֟Z[ U@72v3ydUIbag#3 ,jv15*/. DXZ}.~[ߊ 7Y([[r,cd9LUNTY'9:}e r'H !#av೫)TS*9e/96:EYfVf>2\'^BbBs*S  [w%HrҷDZr-+I :7 \֟;ArQ)5ʪ]IYe;*4#33#3㐓3,P51jx-ET嵥ŒcPU]_99@&V,hY\Y! P B l()A̫ae izy;7F';dovլd;1•GRj%Zxj,~~f(<'dBY3<]d:a4BLO `` \FM&ЬYfz;pEzzأՋ ^f?0zbW/Fph+1xrઙc?j2Jf:F ȓf kcf3\!\\Lf\NTYkqpլt<•Ud'WV9Pn2pf*pe[w 7vjV̮H+rS+1ء \5sIO^h#5•MJM'3n*pլ57+gvup-YyǞ(胆G5E1&<r'An̏7a[M|>_,NsSIT?﷍I+<(ᰳ ݾbp HJ2+ʲ#;HvIlu]U>Vy@0QJK;͸J, CY-dhׅ#7#p[{qH*-C뺴5)lR~۔6MiRxJ/?j*":(\;JMz~ ;GC fi#@6 )GChj+{{(SDԓ)c#CI'Y(` d{[I5<³`7o5P4H!Ӻ֓d,ᘷ,y2ϟf7kU <@~`T2MM!>Gڇ1M鵵wK&%YQkn9Njl?S~g׎hvV<簖Y=z"P:<[ĺMתT;{Jsꭦ}mMoz}l-xדc.gyswl%絅2=qYw>(^ջ-TwOKwO[hc[9m75ȁ3%Okb='iL7r!?7xCv|uO֛:7*>JUv3?O]oz㿳w@=ʖk FoC?5[鯇C?] [A~]Kb>Me@,3 Ň4L2iف,2íji0&֏ AODM`X'JHs6idgTu<Ҡ14I&o{|:duwN\&pOr4BrrPĮ,W $M &u gw5bBh=1Xs B0rGFƉ!Y]2b#Y00dMMFMǍu>.T Yo㭓!@+TEbvvrۜeM-~&7;w+ChOjTOTTѕ \T1;$S\/(u2vݕ>Qc4ѥ6]8{e_C9-T{adH óF2*,yΑRz|'ye؜u/x7ҝ)(k PdMt&ƃyz%g.}>w֛<OJ7p*w𫷠OV8mZG'Sx󪊥VoivRBКdQŢt`'1W-P Cʦ)C^Al1f[}N1|&b肍dHKysç`TVdlMVO`eqܦCnt>Gֿʕ!%aʩ@VE&C*llE Ph]FQ_XɍsZ=gw'x}K'nEyQf[E+JG'-N9;]%u]ћhQVۧ<]%(boћ{o`p:'/ޜ_,4R.ܻ% <͠= qb5ǰ|*VOjX "Ď"i)EAY1DlwȑH:.BTu.c+h_zGFZ$ TM(l 80Ƅaiͼ|+/׎ Ky^\QӁ p'j-T]!J񮚠Bm'msEbնN5hStٷ ĝa7q,vSixDwS4}%J T$UT-c9aN]ҪP"%^,F*dHS,FƶBSvg(ok/Ts>H/ڐMOܐJzlH9Vq=\c6 Z!aa[- UqR=c7qJy4P_y߸q/zPC7b MdY s`\bA)hpb/80OnJ7e*s^{GqQ}H)Q.gxV?|iaQ! 2eاӖoĺ{n0 `m^?v.X MXlVov|>\qB#u%r X8RvS< < =;Mp"zHk&Y 'G*"LٰT-PUW7NjZB\gLF1CZU.%%XgV3,ѽ4cU˖U"k캥2k < U UkQ ٪Pav&%db0R~g.S) Ժ2`s9!KֺQPafrT;۷8{f%| Ad)e}̑*C MT)gmg[3uXY=b)6[* .tB^kH婨gJ82ȗe6^8@l#ASr9e VgxH"EI#cˤuUuAh2Xb)'dG(pj~:,*O_ ןFuqlIL" $ ArAH61bhMK={YnIJ ʠ^D@)~ u .C'C\s"<`d kOH\.3?μWma!Si pQQ㪃i)4Ok8˦˦p6~^nr,u t28K: .O.U L6YRG8/ ,ZV;*${lX,gn89޼=}}۟N/?~{:=/?,8ߣ-$0DM#>cp44Mۧifߧ]f{uLV|-@dͿTWrqЄ|Ig _+UgHLS7ٰ?-- Уen|7;37dVNj߷o\/?,_:+ar*D6gDZ҂GZҦdqMݳ5p!0OJH (A& 8ϹSd(B8[韱5*ADiGjd>} wE%3]bmn=5\o$BJi!թ9:N[tʛ>@8)b+3teUn:{i#1#?4g 1,%,,gɨ &)<@rFϵ,wJ`4aL}$wݝ|˾2b8;Euva}!>. 75qa.*47v[˓ u :tl;*6$#_ YڄPd3( *]Qm \{YmsX/-ʝDu7,7_TcWBbqݸ&n1#AKpZJC6>9&I(Bv/&))QmGs}E8#aREbD h$-+Z#f t߈-iv(qHMnJvn²^E4d>zXLJEJ3iCZu%mP ;!JF!0maei-n1E8s<iD` qUL:[nP.ɞ0K}~^>;!ۀI:q(cI'R=7oŔ F,J2()veZҸ՘IDk1hFs+ڔ[#gO:7 dE10Up!x\6v}pyoIjRH:^7T]W=̪:aԶilo"wV4-q>1In]u* bIAJWOEp .ghwfU!̡d^5Wj^ZX٣浒CJqnUYm痵PquM.zri6ov}ުcE.P@*!֫H()u|X/k9`I&!csII*I%uvȒ-!81xQ *Rz5ђk)SK\z8| `q9 3ժf0bQIkY+냉KM-a$EV 15fbj'աtH*;6-ڄ@ q2k')hʋoegK:ڢs9}џڒT#V{>(E!C:hGXdZ 2hMAgGAeC O O@#1X@N@UBdxGv`2.%T8 y JKI IXZYh[Ç֮*`UyiJ?ɻ"CuVaYѩq:􏯙|dQQhlTlAĒ+<ǀ6cUH"n8Z7C=<гzz֐–b8v "-H<:GMpH(,S>=Iq.yt8^<1rt~[g GA1Ñű4v?!I҆O1*omd枃fd-}6 Dԭ%g"[BW7*uq@"4sa0~xf2x6fnrwV¸t|a_^/bh0)r0a@y& D0@b6 WiJ[Z:J([__ )'+ :Et/j|Ӕyb<$AVYfO^t4+(XUr3TRDȲ]$'kSxAa+X:"uE/;/ҰW_)L.NY7GCuA`- ArJTNZ|FacVR-󺺿eO-? w6&t#kr{Tr)pVi:SQŶxeҁvqPKgzMv{kzM1꫋dWb߰[v~rDjBy) $y>iKdkzM\0ՍBĸdݼڤ;]KX=% ark@fEr|#LZU6`ޜ/]YN^f?lx<(H7g i|>/:ֶ)9<`juݰ-{˜a;[q(Lش`:KtW`:QաtR0aP.q@0a3\%PK7WjϩWGkş^gw+5j!v'b|4pu;*:,R+] ʉ\% \%r5 \CD]1p(pELw`Fxg*KUW*Q+JTJ3+*VCpE":W@+p*.hmp|IIq+ X讘TwFwJJRp +‰\OhW@-]\s+B)wI ,;W\"WZvD嚋TWF u @\u\UP9Y•kY#a= l"blڷYoƫz>jg)YrЌM[q^#X a LE6.` k!>GpdyշشMVꍋ)Е*|V~-\(cV~V(K w͢9['w__,1 r y$,H>̂l:3Ӯ̱+$D-S7N^֒O Ys p5VaJ`Ise9]myj a"DK fi!ӬBMpld[Z,-}zݙoY$X=1ɍt0-RNudʮ3J;d8L+42&΂+UxR7\!!p4Rq\Am~Rw|pȑ W*84\kd\ZbPU]#pWW*,R֎+Lv9SLJyAugeZ;Te؊sUDJO+ȵ\T.WJ7\:3]4R΂+Ub*W)gsaΰ[hś0=OiUOxUm\% m>CL'gʼnpфipri WCUzpUP{xR8Z(M8nr)rT-)pX[sMTCq3DLxC=Zq*l:C\Y6D4 ؒW*Ypbqe Wg+ǞN}RWPȬW Wg+(2]\I j_=Tl9ʫWiJy֮rz֎+UIR9*x5py\0ڕMv\̝uJv&\fuiR~[:K\%5(\X OnE$M̫>R`iwoғg܂>iAHEXk>faFsV|nw6m9B Yw? ,Sв,LXS͞_A!ZZPn{$TmX҂i[Z8åDW*8ͳr:uKU+zs#d,C<1"dNtyJV+Z+pء' ljpi\\gǕ`U? 84 TIOV)Ńt=ݘaK ,]bP< UӪ23t2d I讀g2 Tm *ݪΓp4\}nȐ?%E?5M'~&"{_?e*ye\ Wz"L`4 TnHJզv\Ae$pu$B~"\4 7iܕv\J9Yr_Tlfv\d7\!ز34 x&\ypj֎+Ŭ3ĕ]W,O+y\ڴb*npu> Nw2L+4\1Ӭ]Z+kǕdpuδvW*7M㮠֒_;T֮W)yo3  >YJw?z+?bQ՜b} ߎDHEa⺏DXsr=-~;akȮE߆s3fp'Z hYZPά}iAU:9.-$w gӱ~\ZYJqWC!~"r]|[Oe*ʶŸr;ޅ8 W*y\ZkǕ\pUpe9'Wx*pr}WvR6\!& &•v\JUF qޓu J2͂+U~\;1D; X8N+YjHfRv[:G\@QD Jipj[;T Wg+}̴v4`YjW֎+U,q]+vf܅O$ɰѱuA);M֫7}~*C|. Ϡ:7\MvT7B쵠banI\EBR['>Nv^|C-}7/G7_W^>u~߮9:]j [>.o>_J/~>o>?rzkGa&]B7Tb|rɯͷ~Im vfr4Ūv[UOC8C?!&gEi:$kTZږWɢD7oOoZ޻U}ϳ奶$i~՗7o/^~u? 澻FӾ o4տsPӧ C -?+n\o~"tkj%w/wo]]#Xn(_vgٰ__&W~>9:P ?70L_zqN̈́}M>{޷Ya?~6 0>UoN_{~aa>@Q;r`3pE ~}yk  5z?=_ua5}^7A/o_fEI=[#qt ѵB$ΔN3J1'?ݍ$a7Woܟ_2Sgf`w~lorvd$%qnSR6.p S61d S)` ף>?"ɖ\bw".\36뭗\J9;؈r}_`(l֡;qAÅK! L˜D Xg)$ď|C2&(1W;t@X(/d*هX ÿ-x"RYr!=%ΪJbzl5!8g:RI0fJj :澷1yX)Dɔ{5nj9o3Z|1 |Hm S jH)5ͱ_2"$U.s֭OB6&0-b>Q wԼX].i#U![3LFr.y8iفED k`]ң&qC"XdGhϝPi٥ LdA,.TXCbbY r30أY["kaF^vÎ#z89jD8Q\zpyUJhڕȠ-]^0-#1q)rH,&޲е=;J#` YZ{B0q쾃b$l@6lmbiC .VD@5ziٻ8Kohyd[!brLnAkuWH%rĆ+MdC2#4!Lb+(FBťEn(pNZ25XQcymF29(\`? 5FnU{Cݥ=gLAA[@tLnpѾ@S<T*1a Rxb΂# ݄cb00s1;KI*)ܙ-HFhUl 34(נJo07XQ;7)4GRF;(U BA=F(Ez@?P֑~:8傂T*b;F`J*ѧ^[ȺWrd0Wz,觊 9icͰȲΆ(0n4l]9>TTMw!M>ldPgF=X0׺~Ww 4tlpBrr|Z1f rbZ;L'"B`BK>_m@ndEMK8r Vl%V -XFX;K\?r{q:EjC$*#5]%/I% g0Žk`)`1 E">(Z%:HtWuUww kҘW7rNqW~Io.Ƴi\""Y‚QuU  5xasT`->R;5Y:Z kH繨) 8SV >vq=(瓀'|A _𙻶 J%fM!u>f戇pCCK6s A\ ECeLA(<" U@%5=N(x` ;`X7klUp>W b^D W}\!N..\1;9vҠ&nep ;bۡE 1RŢ45P {ֳZ2 ?y͵aQЏ'#b "KkLnO@jXz |,O>9.J3)%LTH#jJҸh 'k廟kP*mA6tBe_j;Jh)dF)]xñli(1<⌞TV_=x"`~V;͠c6U˥7`.ń12 3+Ms.BDaCdmbɡ'<"a .-FVXX3{ aש^-ilJsühkH!uH)ejLkx7FON߿}kJB\Ujk(}>7-7}?],9| p&ċѪIoꙿq.xͦ!7680[,qrq~vƅ}=vhi_8[;-'}(q؞~̟oԧn,>OTa2FN>PV-sP?q{.??qtFC؜NYNɔl w$TTU@OS D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$%Z OF q@MuGOAJaz$>y""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "z$|?N֟ q=q#ѓ@RI"HIad"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""k#ԞTEKڛ'%/M=N\^zV!'rEG@HI H D$З"ԁԂzE$3?$99 @D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$б@j8ϩL7o-մO/?vRbw?J LiO \mt%ɀKmf=)%'p5KKWB  N\5q;sդGY6)$s ͕NiqB k%N\5q`-c7WMJIXk4Wp:%s0'cd&n +4W !s'cdU׮ޒzʱFn~+k /Z2z(wŘ=}קĭZ8^D'Ԉs-};ï۟v4-Wf< :gq17bs!xl+o>}x5Wo(h9~urrPj`r? 蘏 KMu cwb9NEIv;<߲雧4nE lR~[ TMJ]ռ{Zø~^Z@RtC[-痓v9)]&0/9a->饮Uh߲68BMŧijvu|yJM:qvmR+D>囆Y_Y' t>ڵ$OYSzvtg 'UWkUt ݰfiݹA,ktVcWE,wj%DŷbOd,u|e]jeȺUTb:ztz'x )Vn:co55ğǘ۳*|س;j7PzQ]ItK˪hD,bT^M&8nϙyWmWr?d؆wڰU-}}?ŝd{K$r]$vnj1qɺ~?7m2o>fgg1'KtVԜWF[faZhhe EV%([l#d_6~s+v#Vk_X6(Z\YYFnl6tj/1 /-Vh5mF7N1ݼ.lrOaCX^ls{Z2|ignS9|KGhzjǗ\hK[;nzl2Tyvڤ/[ݧIQ 6Go0FϤ2kt1i`{}}Lzv\z:HυKYVXnADmZ#<`z6Dul0qԳ^~ Ԭw T%tiXZU;DV֊E!|il[M Mœf"L&I*0 |ZPa_x'˪ z^Xr"e1cۜ 'XEHsPR'N`VZN(N8iՏb7z h_㛕G1"uۍYUiJ5G*섎*L ib;=3jyIcñ(˫'~c.,iˍ[W 6kk%a_Nʺ%k]=}i^JmnKVY%ъJʢxʷS̙;oT$mnE] СǢ| C |f|{-SXL:,: sepg9!` űK6ƷFH=S3֥cюFȡaAmbMarӻoKn}Ɠ؛h,?f={sMЖ[hZ >kQkz;0N tj4qp,V<7tI) X 6WC4-֩:]o 9L%Mb]Hv%ZHssOA̹\6(}TS6 ș b:+5W:spP<NnMio|^Y:l9bs)8eZ:m'je:3oNJ4 q`߀ 28UFrHɴ𓃬J_A{e1{d ۑ}܊r,ϡ,sģa4_Ji{k ;}p9xLWRF-*(n; eɩVdc/;]`l 6@mw n mGx}'V[;ks]ؽ6oIy.<ٝu^=c9=|\{qoS\W_c\kEj2p>;S^Rtڿ] WXD<1X[g\9`#{W;].G3K{YIݹ?]J iv5ɽeԒEe:Z6U~ČK >4roS_W}'WY4m+ʀDw(i^ϫR )]`G.Bw韓Ooɟ|`Lo:^< !Gn3ܶUph{_kdKxZiyXgea.F3:5! XTc%pXbcGXT\X[y~!Onanǽd6!uI I'b2 ˱(Rk.!IJfUh7UUN%9$t-X&xh0q^3mhIWI6[`d^._m kw߮_r[y_ r|y= wy!ׇ;i_dZq\х}äKܢOQbD EvSNJ&]5o%SQJbls)i`EL=wދ[ 99rqrW2濐MO.SL?Ř\hNo^Cߩ(CZ=J"Fkr9l[vA+Z*.|Wバ(Qİua5 aZIm\ \rȡdW dÍ k*Жq0q*QCCmPB-|V[my<ŠVo pN"-'uA&x-+ Yt5XWщ UI0eK4VaH'ch| &clX`6ɨhTg[1P(oj*%ލJ-vMVy,V{qRΡ;+d_9! &Rm-TA]4)9»l4Zhʰ6uB ϐ+DI [ KFᒁBUgT48>o|K3bZā# K,}#J^j`ճ>C("_8tەPE55Wav;Θ qJ+mHe_6fֱξ@l#AS"<$+A[=!E;c6鮾($^2:R,iP!LRخItM85F>NW_:&]gc|Q5U[ )쉂g,R@\Qz#@PJ|q|hpbSP7X,&ؚ@wI.ȝB5nט>fFp3Y?RJkѸ+yoq*(~l RD06ٜEKrGX%5ܓ$=n3'ѷң0THNSkBH0;##Oa\8cS&m/m 26 #z!Uxd!RQ0+Ca\9;Aqp4|wL1>7]t6n6Z5BK];,Syn%c.,%`*n!1h'DkEOlQph5Z`(3*h2:ڀmT 5CMߡ4Fg}x4Ldg_ڿ,P)J s\J$Y2$j\,X1x2;-ƙ\N!e :j˙S:WstDpm}EHqt":"R6[%S yD876zE^8NU1ǰ ЩLI߬!ca@wsš%DZm=3{ 0 /hlj)rЊ/|zXÅK/xS:RuuC^ cb3]0e8N%l:Gbb~˞_XCzq xخMTxGPIE$s2JaQiƝVc ;"APvdb>({~,3?eYmWo]T2gD8yOAj4IV1*  L)w#]ϳǖG#h )GpHw",GĐ4%k7RI@RH` ?PӇud1JRn= omg.L2Q)N,~(o \z :HC 8iE_&ٿ#q&,O +jCBgʺqë0/_y}ן޿x&//޽y0R~5$0D_/Mϻ7qSMC{uT]fM]ڕMvyMdnKn DytmwWǰ(ZO"8+B|1} u3WU*FcDIZ7 UU7%H_YKU89;QiSNKy'OTS[jUݛ򟿥\ʸtAj弼zW~uq9 xFg.x&!w/a"pN"rswVd޿LJײ ȅg(k)U8&[9]R\:Js%9 h_ʒo/ꇗժ0]kWE8V(k}8@d 4& ?=m>olz IS˺Kb9Cu}z֧r^3cS=Snrg:f:'zВovtf扖nkE7ļ׾ xFۊ9Ճ i :<&ϲtQZ~-R_\r %hvOteJ"άHi9eTﶕQV^ߏQ]qy,+'<ȩAiaJ  k]Ko(1FqL5ĸ|<$?=S W:WFyl5!7Y?\\.rX̵Q*g er+gI1iZ'uJgC`J$Oyj ,-'sT"ym4L%LG|k6%P0" (t@Hm'$˦]~oác.ۮtϡ`q*8iM0ux{D=x8XJ[5舟7Б> 1,@*z:sK!Shk$|6$Qs-9˝R&$} H"kKtu}2mwrMU-N|,\WT;2am08DGvAFcyrn>?5ʝDW*sPHwףݖ׍ 0qMg^ƒ䓴NM{ۆcUW*'j Z8FƣT "N?.#S^R*ơ, A9\`F H0ES6ѧ8Qb63|CbRrIXK v`ȦGb"3\($P;8M(rn#"Gw9lVJHH!MNi 7F4nIA6m .DpIv<2h!3AL0g*rVng Krϙt:ܠ\.am^m-m@Ɠ8 MB:NT{n XS2(:o30{-#chnD4 /#gG{*W֏OG}E W얪Sxˊ;'r#iw0ˇ&UZ? ]Lo >ΡtT:/Ko?R̖ ]y4 d%j^}Pʫk;ԼPrzHIW?0.k~Џ9:Cnt/~lCzliݯ&&ie3/sH\b+?KRBrkegͅ 9ʞ+&V=sLǑKD}ݜDd q8mV2 ?uSДoegK:}e4z>\##V{>ݨ`#BuЎȴixeКF22qZqZ 0HkpG\84 DQ% H24؀: Ʌַ Z7iJoƀn塺յJl[OQ]8Lv(46^* 6Bʠ@bI cUHEpJNo Z &We=kYOy5C^aK"-H<:GMpH(,S>=o9O#'uzi_FmYq̜V*À(6 3&r88&'D9i\I:Fs4جb͞4Fԣ@QӒ]n'HܝȆRL\$_ޗ6ߦQ!*Bm/sEtY14Zlhn9 0?3MUu3pxg*uѰwGrz Gݛҫ{&r1 N;8*2~wˤrLyXWӬrfA=hB{F>9ct3l/6rdf(p'+Vx"8=՘SS*P?{g㸍(, }LIgvP$Um]]{QՖŶr8Y>|z$iλyfg-32_^QC/8 7ͣ=!٘s y8)T42's1hc%c{W6=lZu[cmcݪ ך8mJ [Ϫ7x(F瓺 !@W?D\n- wŵF[ys3ʬ:j''y>%#/ϐmzdҾȪE6:<~m^J8F/mfR]YPԃEA JQS#hY:\PN[m.,"h Cq%Y0hæjcr>o^w-^3Rn}Io'ʅJFRhBL{!/(#(A 2ְ>"ts7r>M=ExO!\V!a^@W&2Y2/r͝9"|nJ ,Uªł*#qyj p9 8#a DaFW@a0]*wWefǾZ5I+&ŰI] ,BRبřx vӢc-cqbNzW"dm,3AtQjeg#76hbzY ky湷O*Q~7/C -S\ms{ŝ4%!uD5Q^{)QQ$ *OXSZ+}nwvn؀E%׻ۖ?X%Y&W;g(cr(ˣLd8WB/zi5DSBյo7kg>FI{\])B:utU7yV}nT3+FLQe:1w y_?[=v+8I |·ڻn}f6c|{C,}Mc+N,9/vaa֬D;9Q]D 2˵,gSV\] *;hO$6N0 ΅9 gqfU)0&eVřpfὺhEF:E* r۬`0:γy³::ab~^ߥpS'R,sDЧJPE}e2[qŵ^^ƻm*7(z_qo X._z \~psS/fqg6o_m =1<ⳏ6{7s=nҕES`=ݫ#|gDᦚҗ>*sA. ~zUo<ձ4GP9w吭5JXz|[B a+NʵɼӨ_ ՖUpqUKȸ'٪)ެJѲx3UWծUOF$+0P6  pu:b\K@0=zxl=B+Tuq*NWOkJZc]\3j ku* TDlұ@.TpjoVOSĕFi@V*\t+Pk?w*ĕb`iw-f PmNWI!R@$ dpjeQNW0SهM2#jȀ6f[#TKzpTfͽ7sO>78,6y+˿b~Mqƶa.Ɠ|ѿ (2OfTxls^9lu{8c殝_KRN(M!B{^fW]CT_(F3+mA2y gUf](2<3!^MVr}t,R %"OG^-pֵ}ZV26g+W=Zfj*3XdyY&8 RddL8LjÌHj$3rNejՊֿgFRuS '8`9P&+AS*&Fն?U͗+e>Qut\zr͑R NqpUO%mYX+jת\qi%+{аpjձRW'+-x+ Sʕ:\ڗWҚW'+.6!\`5 ԪcW'+!$#4!\/'W%+TkZ JNxĕٔ@0W2\\-Rv:E\)I S3ri2oQ-k)JKLW X PAO W T*;\ ᒈoP7))WȷlyIMcj}PDwՙCujnC*VQt 7 ,5ܤb`PX ,ԚؠvA_mz'qÏZ?6=Ԫ#᪞Jݲl \WV=e3v_\`PP *v\J-;\ 3JȄpm:Tpjj;P%WNWxXg % PB)Jp46!\`3("2+PKy@epu\In9IDB%++m*BBVt:E\ #]B{5O&clhHC7ɴ NF7PK)ec]\zw Õwg~v4/E%vMp] (wgυ[|6D"[ 5l$,2N@܌罋8_~{$졠Aoޣ=eg΀1%]Y~;^us*9եhGeki`mӒoݟzNaj8F(,AЄiqw_fFUX#;lufJG?]/T{5Ֆ=8E?.L!)2 j Xtl5˒yψjo*UZTJ$+)OW(TpjE]KT)yĕV"bڴ^fep6ϴEtn\~_#}$iM W AV'i+6!\`cʵɼ:W`TE}9dO<\,RS<\S>Ά5UvYWlŽvz*"!\T: Bێ+TY pu:bbePsbc;#U\O"@IJ~a4($ e k7a/Kgpe~I?=u춸WW,gg!^`d*JT`LYH,҇iVU^^D/_pl]\}^P2e':Po)q,y$by^ͽdZtGj 0gs.k_t"PbGoYٸ)vZ+ylWR.zUb_}.]v{tMx왇̗T|Ƶꦫ|wՖ[Z;/u7V*B )nO^n/Z+Jcޝ]<( C? )xyf2g9g.S0 V7zR7ތ~ |<_J o'?"Kו;[<Cn4#M n ~99~2Cu2J=`gjAzӛEHΞ*ui.ZtM\"XDisR\R=Ѹ٥Qܬu}ėw]p vx{>٭ j/1T+б ܞOf Zs&5)+HNuNsV8AgTr{d9XX/sǮ`'UĖf ^ YH4;{Zӛ B" =6+XPtv{y/2P[Oh>܎K;j to)8ðG?|fpٌCޫǣGo~T| ɥ:onJ # BN)/XIr\zA xv{/'_Bm2f_.!ӻ>gB@=aCzߍC7Z8)$ލF*/k ?h- Ӿ31Uϯg0tt`{NdLnP2ÄY55h1c5zp T/.l?-wJ[0w!eN=,gM߸gHL[ yw3N5,QM9YFP.qX^6- *cL8gYF ̕'WWC{5`^cBKG2}e4nܟwa1Q.,~ ˜R, 3a8}ya:<{ǍJc_6e/ lb{7syp,֬G3s-#,E2%փ=3&٬_b8|rc nWv%M]JID 1.`PꠈD{oN,jb48p*gn E*J u.F%3xCm/{{K%U'S.yg+*L ,=bߘ*'S'?ԓ%P2znK>`ឮhp7Th 8ƴHH$) e>rX DsJ .0tY ."F yB HLgI QxT͸Ơ1e1 +j1q+j|7mXeIm E@{u7nDތ6 }l(~®d/ѱ u߻SNӻa&5k<&D捎hyGa| o%ߝA۾P\iɴK^rbHbIT*p^z!1xGH$TW3<J n #SHh-L%ʬbN P¼8;ANz2:J6t2|脸 ]zR}jeV')e#ܣeL$PuqEr 9Wc)ZT6)0j0 R$o*ZJȅ3Jp#3):5cwT H5blkyWxDp\\{uϏ٠ަh-vEoY>e⬸e@q9={ڼpǞ0!ud62d-  ˽F+ ΄4{dB0!Ho&E- Vo iPPք1A(@%XHXIk}NTSN8ATzi8 hD9o2?,\^vXp cx|NUe…͇ÞƖj3L`ڄX$Wb&XKh@qPZ pM.mf azq*)?J.IU |=* hh;nbocjn%߽-vqatDq0?:m8.'8=ZkO:[>3*1ezbxƧ\/u@찄Dj)%U2 /KŲt3)퓃—19dr fgf)ޯ-XٻOL{%և|>"eF L$amj!r N!n@b4Ix;M,E#iF)@D }$sk}1p==kV10C&'U5L]_[Xbs5 k$«uh$e 8SZG wN)Qn7i"$; =t~zY ==|mØPϐ8IA Ad*ڣţ8w2x9ud717͗ B8XmTTQ/P={VO:^;U-"5v%EQEx3Ջ4ҚOsʚ:\zͿ0pW^ ctܓ@+fxA fzGx2f݅*MVn^r^G0N.nry<\Dz"-Vo/~oXy+Ym˄bx}(w{7H)1TVx04`&'7 X vr#pb{oK`V 'IWݟM*"Ul)H'.ǻ爼,Vy;˯aC۞\<ݬ_w!vr<S]b"bksk7%rulav|² I|]j%\4,E8aC V.fȔnW?1;{g]`<n3J\mq-h\$b3w}tSڭqJfURˡ]y.]_5uWGS# Z ]xSB?/q7jsMyB@a||jgn3`WJof[\t96uX.2!ݚJm[2Wen՘Y;LLf6Cwkcsf=(=t=9`9. nϗ#f6eO W3:06pK%i,Qe#~2ٞ/Ψb[ox՜0Z#Fϩ*h4pT4ғ`^^)4md,&AH\Yb?rસ}Âp'4Pʙ tB_[ʼn"=gJk  b܏L~G&YQ/U(ÿ{oEUZ GQdg6@(1LYSB桝gAQ@E"B%)$ kQӖk녥(I!FcJ`ypKirWrDݞx$' fDgoI`@_wWN[W'ʅQ&QB(4D2d 5,MEu(\4u1"' oMpo,3sULZmi ҧ)gʹW)>i5䤺!Xd8$ xA/Y⼧))ܦmMl|eB1erĢT "qjTGϼa:{$LŸ ɹdNxI[h^5j6+c 'g#.YāQ1ř[m4wz#FȎ=}i`[Ƕp3w>;IݸdFi>އ:Z Phd)`{ s7,߄/3.)^(I)IE"ȤsI"!˃a8H !zF99~x(5Rn΍6T`L'jHvkL RcpM"Jf $רthQ\>|v1mnUPZB6T8rvt4"!*.hXk>i|GMi+.?!H!0kmT.gʇRLR<1|+ o?""iC*JQ0E$8hw2{-D;P-zE*H!s$nD,!PחsR)RxLDрґf5Lݣ򇔑c޶;78n*6;Gi|wk5!W2v9h5ˁtk6_)qC)=ոcJPG՛\`r{?)3|~p{`_,¯_&ݛT޲in.MOu.꒏ݫ(;o8GF8(>0 /via8/4d<jٛF [LI2Y]JJȯvTC~jr.Ys KN7lKW9/θ~h,v*>]jED>[9ܰ{Bkun!r={{~vl>N"_B=Oy+}wՒU WyvX貀^)hٕ:wGV)5/[] ]' Tۯk[ٻ6$UuX`nlA`SY&u$[~3$EʢDMƂd󪞮Uu=~hˋ]z z-S~VNP?~QlAج Kj3落7LMЪMee!00>vMmw+n]o\c{jU%{e-( 6G^JL)5́[ S࣒θ(3,OV+5KBeoa=bVz=\<`1+a>aV>9CRK#TeGh2Ć527f_e z QXMl?Aз R'B4N%X0JFr7]#i,mbYN}L $kPot(" .k:+,Z诪8)qDdYĸYdH58N%0 dҚn82NF]ln)}*q#k4+%ʉOnt7u6|>u݆q)W۰ }v=8l(w;+&Ϋ7ug)A@b|~V$|Ң"$5 (>YN i$AGuңےoғ[W0]7Wk4ȓ pߞ_? E}Mz4\yey,cj}1o%zX568}payxK垢rݫm0=s#S]4YVs՜UD=%8uRCgTk0!UNچ NYq*s3Q&:^A D皢v9HA;n6=a=XJg=*%+empphJn(00Fz\Sds9:Tp@8)Ʊ@jMٮi]MWOneWQ">[RF$..i;rmG1vmڊJ<䢥L)mS="g횹DWw7SE^/?$}>Vѱ$tz2[}nŠսkf$c7ZZ07wexz˓隋p)cz۟~-_v<ۡ.DGočxtEzy<ͧ[ğ.Q_sO"T]&(Dh<؀ ʱF'eJ<j [C;NC|}{D,gNx+A"F%HݦÐ89OpU{ߤR* d著18-H\d5Y6_5;z>.;bZ1E,?_w"tF ~l-Ltr>?@Ia.'8Vaҙ=ʊߺ+i.Z|S>xZ![8;9ޘZǨ1kGVh{L@m]Bj+{A7*ЀLxD4,OwҷwH^=cw%2»Uҥ1'znQ Iܘd#n'Mh<]c$e>Y!H`y) 6}WI-rUESԊ֗ӜV ZhHE/=j`I^}(>I 6n$ɦe&I# 73@$` zɭ&mcHp/gG [\Z2c9ĤC 1XY6\IMĪrXv) sbJa"1B"8cJI2gee9FΚr p&Y0*ޗk%I3: 23{!lqFK'3IFfID "pR԰,֚'P"O3L`'!w:6jUZ⬷z;V} F/T%¨\V -@]i@ Bj,H>he8(['^H`Ld!uIi,#5ApGHL+ >FV省Gh+KH+Md_l*|0 P' 2Td 6fFp֔2X\BUZ[cJz's06صQC6SY:kt%luA.:iOWz[ QPf }3hfwI LGT E : lns2MŨ(i(-"%&fqpL0ݰaO,=gp28(w%P(cKJmzC/J:&%Gcy 9kϵ>9[ Fskmå,N1 p'|=7k%s$nPyuDyDH^&1SѠuVگya0g]9v0²O"y4D_+-m ";eAv&>K'y&؈v:~/,ueBpe:+ ~#i:1e.5+ b2g'S!4=JUC&xӖ>ݪBj ogze3%-qgr2mSw2W|nZ=`Ⱃt|VI)oYXsSJL 5-^TeF_u+j1$^hM;2|2 ୍a1@ c#Z!T(t?Vؑ ;C~;'G85{gOO>N]QSTrdup=uȕXr*$م*B@gn=z2!wJ 1橺+@3a]llAڈ|ݪu_7'[g?usf ~C&|0ΘopvkOP 9*>qǬhCD!'#b'hUc ~rO $N>o/ $Ȏk3ҽyթMnԛ7Ae]j{']tJ-pqU{RRR'ثLEdwu_//!7u=Uޱm'vh˱t;.b7N~vm3WǵH{r6[^X :ܥQsLu8I^p>1Qw 2{Ø6BJ0@x}uMw:N+5 yI5"^Ò}Dsqu> &p653+Z&ƹ XgzM в}aiqsphGhϧKc+Hs% ( 8c`g7jzєx6w|J ޝXS1)VKM]]K VQ WKdA0Bʰf]~o&<;XLXך{iU%knGݧh+Mͳ=b<JBP5bz"pԩ5E%Nγi<0϶yHw T`zK\UAA/!6lL W#ഺb)5f2/8_9g95U J snܴPԯTkU':镴7QBiZ`cȔ*D&C̒ b4G*q)i?JGic:?π2MV@9k7.Y,8ܺҝ%(["1 P b*66&"Ѯ;&hnAZ H k¥ZO1Z2-|8((E͕ .O[bэ+u]U4N$Y&/dј($Dϭa@B* Lcj ڔ&ﵘOgRhzH mõNɔJ[fGƻy6{5^0$55U|s$ WW4KTvη+RmY{cb*I )ONinl?e4Z吒%&0y WYԧK9@=Lw=r0q5Y$YqIuyƤ928e')RteoL{9)RĕMb|#/.4/:9m_zE3l--XY^ha=C[>ym!=WO7@[/>:Jŕ<ʐJ)5;*Zx4VB6`ǰ{׿ONΩD{'ەi ޘ֍֟._[&D\Eb%{nJBg'z6_|?Q0+7J󮥒Sܶ&H8Z)4P{Jϵ5L< {9QwM땿pKk}i.Zl)G2+duMR"IwS M's=ףWq%G+nn~ꛗ7Fu>6*ybx-,ڜw(}>F6'JOώ?nx;[cz?;Y%,ۿ u)0\0܋0l0uƔ-u\JҖE3+PQ䚻1&XUSXJ; 48-BDb㦅sI=nK/vV@m~,.ϧٮ&\}\ )(Ϟ\p}d1Nw*|ȗ|U Ahq1l PJa?n @F52CrCCG,\LH%3h$UH<bkZN=#aaXX"43^=w$o21c هԳ@)0Z!:k7ȥdCuc;ﺚ328W5È=ϖhKN~ܪzm $)11% D99zr9M"fJ},KB#"58Nj*>5 ;;0Ğq Eߤf̨$CP*vvJPܵuHw#fs?bzMb jOENYYQOaŻԗf42&cC`Nr7,JIɻ(9%qu2lЃh kjsC وaGEpuj!xW%"d׀+"x׈hD6M~]0gn[irSE;#LsW; y2i 3XB8#U` uOm93-;xr\9m<i2.Ҋ+.2A{z!ztKRrZH:|Ძ R^qqx0n~|3?Y:^y㗟WnN?.w P3Ee.aiqɵ M_ T Q rK2 c(z"ԚWƞK+>eeIPQZv5ױOĚ3DRCѓTFhK%ИbT]VGd^=-A 7w`3ѧ/Aeo_rC|mЍu/b4|/% 5XڐkFn*؅[Q贖[z 2פZ )pɭxKa-S+Iti^:i#4GfL g<>;==ﻼjkϯYPޮ).n]*R9ӕ noϾm?[>LBi!KB6ej#/ I.!IH: I!!LI(;^iXfXz ֬}pB@FNk:WBΡTOX7vW}Ң(5 $m IXpgo ~zˇD9Yr1znn;\nXO1/Dv1 !4vw؛¿:Y qRq闏5G+gׯH+ e/9\9R[Iim㷻\ -쁤ëUK S!Gq[c&IXma%L):кϳ}tH,Xp?ngǺa}_L^Z**ûoԝH1L2hMs)J+C)uAń9:{A1/ +001'$GW4iłAF]C0Mc}k派(UVwvؽc㗍1y}wmîa.;h].yyZ࡞ĉ=gFX"@UX<$а(q I7a8vó?ե =_tw}0Ε~`yN0F;ge?P8{~qvʨ'!^s~M:@ +'NO?bCsäY,yR(A nVZU@Y@ǒW>})`g ?3IC}"Qo'QYOV=ӳx0&E&23Ό)x[.x`]o"BljSy6g:׼dgK`oƕ̛]Ue|fһ줟4YA}<$uK,u{(Q;m~dFoMY1+)첮;Ή'? g !u/M2 6"dqi`VWmo W:6|bzݚ^SMLbK R ~ZίDv~Kv9"tjn8MX|I.͙^Eྺ^kC>4ͧ7)R/07ٻ4՜Yf:-aa ˷2+ŖYtr <;x^p~t)VduKim7K9Oo]ZN޴<_fK&f.`0(HB!xEڽ%8 AUP_*'D|:oػZ+v(/ԹXz,󉠺=Tk"<-&Դ2(7ۼu(eI +juew^;*g6!G}϶z䚓3v%FX,tt[ 2R E)W[2`TcN&9P#Ɩ6ߋ=3R=;.Z0x SV (T,R /Þ(y 2B8JsBBd9t4l t"/x2ܟ-}`ȶ"_Xu Jɵ>_HZp%ZDjk<ρ6"j% 4ȜI*r͙A&iig=GF,(h)r_XU`G0pg:*.%7 eyHcXDJy9i7$T t/Tj4vCP q`8-RLp7v`ɶ l}U]szsDIAO=LB1\Xy~>&ʈ@K2$4u2J1,&,&Ȣrrb ъxifսu+$6!Ol'>n{ͩEI@(K谗>+r?JmٳNI|V2-N2_@3Q۬Jߌ'~Qo^6 &҈>{~N^dpFπˣZ0neTb׶*IYԫ9sFD{#U=vWAUlE˓HJdpc䈲QJc1R)uDe\1vԿQL}1uiYLй { ö#(JUcL D+T¦3;kfuTxGDJNcR%a̰4N+9!LK@]cw)ݾY7YQ@,Eq[=ZaRкD#4BEA#ITXa8x[o<=8>c681@rK 1$"aɎݍ6&ϠGR ;BOp*DyVI *:bL8x,M߇e?(;8B+\:(gꌟW)A;!'+ZΊdKEBe\:K,_ *ެT }}*י/(/ޟ~}˛^?D788ߓ%$0Dw&e n4?ltEi&M6{]&6ݛG`2mKnx4-kIW<.Y+o>Tn[yn4ƽҌ&rH:7n`Rtz:|yR*ONg~`׃NRlQOB tt>Dܟg3ڙEn嬺z_uq \1 _>IȋL?i8o4 OᗗIZ 9M9W:ʹZ Bլ.D e j0E1EuxoʒMƗma:keT&NjbXaVA?ʣ DPnu-m6u6/;*X;./Ȋ<r/>U ϝI"hN^i˝t*2B~z$B43] 3<Ѳ4nU 1ƕ/An{ mŌƤ[ (QN_b%,R J(0Sa ҕD*8S#; ZpR%9@sGk˸I,>r*D6gDZ҂GZȌ!x~9r#W֭-op֙`{ScEh86DH[7i&9_\ses]ql Jc1F1sɭ`)$#;mRneMC;ܳ3k}S`pjB"d;Hk^M4S ZP0" (t@HmO@H2HysQ Ll|Rklt SD.چ"c >pby5YC}03#*ff)3tc%vhx;b|\aaOJW8L0ǣa &)<@rFϵ,wJ`4a]$w)砮߲O"_f~yvˡ}!>[. 75 a0xl>۱|{͍d1 )\g0 ?cÆ,]mBCٗP.).$A.qX6|(wݰRYcc`Yg^t&uɭN2I_z-⛪ jNQUM^j]`VGX*j2KySKJ8oaTzc!8gޘ/йcm}fpM$9x˜5j(';X|g(})jSU> f\qӺ[TuK72j6TD$?,bF$׽ :*UӀW naQ)T8qЖbb)n.b|U&j*W[LYFץy#qڱvިyVMs%W`t` o45W\v{Ρ=Tӥt)en_&um{g5:$ۼi{MH ~@`-D8Njr(ư=qc'X± $N@D.CWZ •c /3)g٢ S/vd/ɒeaN;"nTLWWN=VoW1W\a U)Zw+}8pJ,L;\%*:+J<(aVB \CD-{a0Q \1 W@0!`*Fwx*Qy]=Io}@p` ȥ \%j_wGa)•`?pmU"pTZs,Gz:p%"B\˃D.9U}DGzpswAd_N2:D0$B-SqWv=+LSo_PeV=R9ʖ/٠Ȩ\ 5qv Eޙe!6RH@yj&L^ʤ ::\]`WϨLrJ͙"ǜ MpۊeV,s1腦eQ^Oe,:u~?=ߊ0eDul'wt{-MM24l.qw-e:<Ƒ$׿B |23"6 a}\l!?-b)R&ّ~Ѥ(H%bͰK P/3#_W玛֌A6v5Q.(?fko˼sb MԦcv;'nyfAЗ DWٳ! ] \Е YPF} QgO<O}apy`hPҁ)]}s+ybapיִ8] ZY; 0ע+a+t%hӕ̪#]QtцJ?7VJ>5y2t cݷJWn ] Z>bPPGB+3!`mmNW YUH;3ngJ ] ZNNWRsnCtfJldJPgIWbEO΁jݑcy-;KӢlT9 p M |H9G5;U)pf`Af kP4BWd[RWR ]nvf+tE<2ճ+Ao}\bP2(]= oomvDst}̏G;9=/6 r^x{MP_x 4m@]_$k@(̯?~z]Y:3C/?ʯrzzUc#woF?޾F.r{`Di{fߴCȱa̿\Wy[ow_rmsJ)]8O 'S#>7YouﻙpDk̔[>nQa￞ͷo!dw܉WoOe/~]XP:#/wgo薅yxBo>?O5gg0?Am8y3+egǫ^A>~~y~f?_.·Gqfj4 zjCa+ۂ!-x)(|ǿDo^>Drq1r%C30P>~KpfL3#xr=& \COl}vm9 7}5fWKekB[) 9|,v:I8[Pڷ~1Φζ|g[hȓB㘦# d"ls|qejcGohsat[D9Y51PL1:FjoluhFqZݖYK} ]Z@no޼;Z9o: qz9} [OJʏe3sƜDh½t<&3q:,s2jss Eǖ73RbB1spOD[3gv}/4\1v3RbG&gk*0&!jh*OϙlyƲr`IcJxdl9 AudF~S.:ޞ6&TuZ:[d+%s->Z90cr2B_ϛ̪furp2fނ@*iv!f8c[u}lNkaOnadk;ŖA\ȾT0Ǒ֑~A_;O2)Kh]~M&VCJ vGKFS%#|օ#D@Ϧ"X',FM%ن>dg Ē`g{n`B`k "J_k ]5v:heX3ҫgl)U( 2-MCf0`9Vg`򀢳|c0Z S.XV;ΆaDpC!V}KWHm!ty%&C8ܚrhu ,&^е0<3`{d7;04R`T^`1 {Pv RK>PP\h2km ,>S\D?56}'CXe`lP& $!$6 d>JɢR0IJV2I:$L , i%ٕ[PmgfH57]iڑ?<1}1 DahK@ k6L7g)*Z]A.ŜU' sǤ< !.A -i<8%B`@yNC,SV `B)VF 2Y1PHq Ls$eEj WY6H( ':+BrEA}u*Rq&-R'BTݫ.Rz(r?Y3Ton0$$+"5ض(=%5dYr!#*DbbۣAv٥#r|hZQ41P4eYȠL!|O YܣA{ۋqiɲb69#94T8hH;L'kQ2_%`L:|@uPe0}J_7#JDkk% Z2C6a1%O Hv9!. %tE\(FR|$QLȼa|ĩ:#./1ZdgxILPc $k(ARGm!@@8ƁU]\UBN ~d}gȈU;o@T5F]6#H b#җ a|秧x|cdɽ0TM4`U>Yb-=!DC`)B^Z@mDebuj W5M \Ӭ4D]I (ʠv]5K qr,-ChWjюa'"8$`[r2 FX,;yUySla5è*g(I\{QR3L bA l5?\U'KS`jCv&f jR+p* L](h`4&6~H_ !B39F9` }aAAxK6Bl܋[ϯe'kH iC6k([.ߣ8}Owl͹Cevy%7wEPd e__p8z%{G>[{w r|tISQ_I/>܇ZN8:zv 0 7z|?'z{yu=/3IGk.哮N/~d{_wo_ oBn~\\NO}[Ϯ{?RS&zH^;vcg#\@l[rA<\f@@ !9:_J@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N ~Ӫu@;. \[m<;.  @o N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R'wbc\r0ٌHf $hHPƤNgb^@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': NgrP['uvCN %'FsN : @RduH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R':ӭ1T臟dz~\6Fwߝ]=@aK%Nv;%ͼGo\;f;5.}ƥ G5KWxCt%p3mY:t!]@mp+VJpt gHW1t 9f XSl6/cHXUY[ywur*Nf?Gٻ޶r%W|ܘKq3 y`z^L d,8qbɑlDٲNwھkxYaA"Я~|Cj:$ΠOJ}u;<]?8:ËSJ`f хf~ ;) ܫ1R7r#2Ԟ&z8spOUj~^Gy}r"規Ż5;ݫ+m6E=7$PoU='oo_h,h@8AkWɕ{uN>p'Eϖ8&2kt|M٣sD`Y4uzvNܡp䜡t㹇F̴ NOYPw5^V9A$d *8%D0W,J~1_M%B9[Zo§_}J2D#ԹLv?~}3F)=!xs%hž{vKK-Mkr)s0[yPT"aObI7.4єj}~R,ܨGؽ}-Ňsܭ9y{/#~q0Nz|?iv̹Xk1ը;LOy6*Mɥn]~{;`Mھ{B_g~iGMxƎalr6T%"̴۷@Ќ'\Ȧs ,k%FhO+RJ"Rzq^vt(K_@4wGdD|'~"- 8b;lS7r1]WFc:[G!g~&Y4cV7+w7mx87'Q:|z[D-;5|էw} bD8jVU=4ZRYbQ5(p,i7_jɧwo}w=[AJr tF[DDU-r(s L*>8wzaX8~YǾC|k-=tw}[]}}69 :X00 T4g4-sJm*b$MRB4bQqA<* ^EkHAƬ1GarYf| \XՈk$ݭhy؈kLc^1X”l/]_ @KؔYtl 3cym% 'pmppxi{H:KX<-KsʶH`W9K_-'DU[lu2LuS=S5Z8OEUwLsdv%P[9sY`ǶQ8Lփe;zv,uTl&Zu") sIpu 5x'y|tx2 1)]^[ hO(a#&"fW=R$9EJO~M\g\9)6Smwn%ЦnTT؀L\cyn3L9;ag|'Y6O3M6̺l/-#zJ;9FDyIKK@.EDJ$&}bUWjTᨸWMi`GkK/Lo(v1KO^>MoKy˩]qڡrÂgfLjLqg@r8L@@ObIXcKΓV݋Vs u.IbXK-K%),ZK3Vj~hgZ̬UN?-mc*h,K:5g(efH¥kV1MI2(O&P_~ kIybhFIC zS C҄|'LJ8'U jwI'0ïugWil "IU"Ìe(%0Ǡ|&?@Bꠜ%<ɓ8FiuXj|C㷚0,%uX,XP oYS6aɲRڕ{Zi4x2zc n:iȮa*['A'2+Kfvm4[.1/7H8=Mlc$lƄfkóM'*N"$A7bn<*Ӥ|+))| EQxBQvEpL0ݰ2ÞXƜgpPlTT'Kt@Pp7%p uRRVcTC!stD(jZlfMiU`42\VS=-u Z ;#Rb`-GWoC$zx9뼓4kl8*:ix(#HEeT)swA!L,!aU?K1ZӮlK|㥘 ]7k<QБgݫqўQ{]Ұ^_]4ft< MA4,A# NfVc՘f5X $2h+FG L>pzĐ@\9E*ŅʣtDXV*lu +j5qHiآxC=ǓzܼJj76\ \Iqw~&J:9n,lL@'YRYU檽aυV[hєW*{UB<+ \8$T㼐NC(bu!&ǘI< 6f21 O ޤylLNg&ÞVf +o  'oXlW\s774|uGmrĭf9R+4(lH"D(n1ie b2ܦ5 M&nǬ\3ګycA?.QcմcW]9ޮ++' XǙFDT"R6Y3JN. Z)ts6e헠1dBEL|ML H¢&H>bNNXMaK1X1Gz9yēGO$लDJ<}>ɶmYd׎"Gm3;Lug;ič]ֆh@gB hF&HXDd|b""QWff\4YMKv_'x"mt=Y VY2n4SF0h%=Hv%ђfP3V!Zڱ?t- los3vܪ6/qŊiEE)%Ecc1*!FΫ@.h_nx4awIOѥCH~;~p=~P &ôGm?i><7I KiҘ좕>5IV WvqoF:z= m:P`Ð|: ]xLk<= M+C#Uf]睟ڻ߶ 3}̇_0=ɰ?:szAO;ӽ/-dr{nŰ|CعL Qr jVhfk{<۹y]yaBԶMLDC#rw6L$5 #?Em(j[wN4jZ~6CF sP2*%dUT3ǕFT&v\8-DS( %KABZ!Lb])F9P7I5gJx8_>]/qBa/s8E %qșfwQ_UWU-Xhm 4wFm6*GT KIw`PGs;YGil?dkO^( bzEg{$C&\btgr9lA@ͣ.gN\b5)6AwiOBW!THu*G(n@LE/)e6Zz;AVE:(@^0%}*hUIy̝v (kPPi0-j0#4ȹVr y˫.gtfяܻ/;]0E mƊ0vALWī|Kʏ=y=k%Is E2zERLN:wL~iL 8iCKȚDpA)C̘GVKQivrb "#Fq" vbJcprozzH7_&G_-}SizV@N;x2vaqڜnxϮƇ3ќ[OIt%YJ\JpB/GWDۿfhNjXJbMna&s>aGqͺ> '\iC4@Q^ 'HW%T@qMF p$%B( aT$ @ZN(74;-p*ࢳM/S5'Ni6u'QJcփqI6E֖U: r#i7 W&5NSn>bؚBJOa8[*l,*9ìvިyQ+'-j^*y=귛-C>oϱžUT\]qa?QKzҬ_>)._-&71Q"' ;Lgidiiwb-2}`ÇnLD6\ [Mb.=iCUn_f^{0;߿\O*8ow.?gD1Wt]ʑ\+h0+ T(˥@ &s,xdއGu) keil(uͨ]Նxwskk/EJr|8m#V{>E!C:hGXdZ^r K{i-W˸ `x5 8#"&#<zDz$l@Fy۹dvh$z 9n6)}?$.Suk+dݞLQ2WV <̵OV(46^* 6Bʠ@cF8&7UL5* NN#xFDO$&y+l ñc \ UGP  %E>p.>=$O+'uiݟ/*u"h8~dJ=cTp$BGb~B%O1*׷ z ̺ -xv[pH`6vkl /!WnV/*\{EAQIt)sY!TpsŘגg0 mH8Ӊ]\[\e0Oե?5{߿*hz-FG#vP1;t20a ?ΦE˯SOjB s~K{o ̙+fNN?bCsäY,yR(A o+5|t(Z}2'iz:JUՋ, 043ާ߳_IԏFqQ{hϦF1KS 귽y  TvU|\v6L^8|p"ohw~MrHY0"+3Ppv2ë8-gڭX~dUu?ptNSLjBN_@֯OY@g$M'7}sc҆vq*識|՚JmW͵*_~~tUhar|Y*-U{>A++R4O`6I~̋[ Ll 0FmYsW4׷.[aROQi~S-zG>It9y?ɗmnw@b< VK7|~׆n Ë=P)ˎn9"f=4Ibv@*gHyePo'TmrTPI![FUykY2$"ڰt]=;g7ݘZ;t 3+EP;iU^;d)ƜjM$s$b#Ɩvvv3RQKm-8 |gɨ|AI{BR,8)"(nנ/S+8x4'$DCseKB/@IήϖƮ۱c'ozY}ӞW?R|5_}͈86<{}VR@sK̙"לtH{qxd$;^<8tբ"'^qi=Sɨ̙tarڥSnЇze8V?IRLCW^o<ҘXtӠ`^䖹yZz41v-]N6jWR`n2_`ga9UFkЮҭJ"άHi9eT]v)޲ӰTQ;DsŁG.Z-VN2eySA%!8$8<ֺ-6Ff;jD"[[;F,.nw}΅>7E?sy)$s*Y̳ҷ cIiFcf +_` ~Yӓ< 2׳Cr4x?;?^'~N&܏\Ip|x^ދDQ-L/ D ( 9s? /g҃7C <(߄i<ޭЪOsr;[|Xnl3^~*>)Wy0;ꉪfWkt6NJۃ?}_):KjueQ2XGYGǣ Y٧K + O1y2*KSWZ]\%*E'DqER=!q̟JTUCW%݉/G\Q%OH\%z2*SW@-UN\}!i=`6(ևS\LJil!Jc1FArɭ`)$"AH׵k {61f[}m2,j81z[q_. +~=[7b sjB"ރ)w;HA^M4S olJa^iGjd>} 10%r@ 9t[1 'qt/jpFWsWP5|>Z'i\0D9'GゑɥX\02B FR`Z#bWHî*r5S UrieW]I(?&LQGî\F&|(S+䡳L]Bvr}LC9ӣaW\A]U2-zjo^?ViըMk^soN<{'h!QzAbUvذpŻ͙ ov9ȋbѶ=>5x7ga䊢 pY-Q4W2~JJK s`D(1{zy?& ~試K& fg*]TSB$gxR;ȕ|iԚ='Q聅'+ٲmҬB)9vɥXUowgWJZv LsD*,ѰL.caWZE]e*5i+dW qO#XѰL9vή2EW] եXq`jѷaޖeEIz3EHv\.ʏR$<1T d)u3,20x6ynV猝p';LVȇ)d ÷?/$[ÉN 9?dڸ]?h>6/3~Y5ZF ZDtbZ DAs3!)_$P` C~<)T8d'd=6:" h2L-C˙JP\~rs6Wx{wxӻzk񑗎3U ei@ҳ7B]GhysFB+Xl$!EB EBVg*8;3@EBzc40iJ~I}t H:Y :4S~D\ D30x Eah"/I窮ˈ蠤P'4h#PFXč@PeLFAFAp3v@gl}kogU٨5-Y{G>1́ڼysMl}(]p̉J+-SVZ &=ph?cۊWaȆ 34% \$b5$5,RfI[B -{ܷ6dvVY, =[:ORh1(;F]ɠS`j-JHր=AJ $:`r (h;VA%"1OhRiU)ukep,I!Dkq^J8ӁINKҲP/,@甑ɇ8*B vTUG)񟷱N]5.;7FMݙ"x*rX?˷?\R/?|s$w&<uBCx -o2jr+}x|$eq+\SW /~ 뫅w~.T/tZf N10yc~3M], 乷1֎wvB_z6@TQl#(L>/Yy~~6/T,xQKgg8ߨ՗mağ"r,9Eŧ(Cr7]O߫uU;q!JQ?.&(hD)]9褁kCwAQdΤ4mcb+Dqp9ӎY]4u˟ooGWkTˆ|]Yܪ~zU U8$Cdž¶j[5 /:v_r19/bdIx_z{~^7d/7OwU@Wl|pl[2 9MAˁFSV̨堝n_U}ZQndvhŻ}ޚ,kXPrIR+'ҊpU6D=E#Gk]6VziSeC4W}w[9jvB.".Fo"FZ'9RVCH7B 趿]YVĂll~PVH“޹ }n@IOaUɉa;Le"Pv\7:  V9 @ ) M|\ԕEV NBik>&, i|e{u}2]gtUv+5ŧ{պI^g@Xk^b"/vS=O-exgmסb.j*ba,XbNbi|gZ'ii^g?9MS*ByS:TvaBK~XO=x%-K5Z0b(1$0=C@}$j9㜉F᥊(qM4'cur:C%QMn 8\8$l&"D9pd_\Q}Y#PSJjl&E!/EsbVxPK)>nuvQ|]ӂrXΆ?P9*^3|,  ғLG!.pV<$Ǚw6tDyOf\o/k 6Lx_'E%̅}rc^P-v|ƜĽnq•YCLɚfw q-:B=zM(tkN?iin'm6/im񩵶5ammc nvҶK0*X󗹛ԼILv7qv7q׺O{Xy^:3ȯ 1(0hzw\j*B( $tvF&AJ,w1q]6 G(3Y\W!RNY߅-zgp|$-jrtV&חSd1Y}:d_}V K_L6/&1SٲZeö>:}^nV1˻V[z]os[7W[ǛhPppkV|9ٔ_䒻&mlZ2}Yޔs~ Xy|uS w~7W;{Y ڭ&5S@3Ŋ<8*%)V%ugl=i]i[,x>k2xiGI7d1}VO3uql頝fv:Ir5LֱF‡#s6vBXkR]KYf*-Wox.‘ v:8i7[GH;K *IH¸&di{90|xmd5، ϰkYβ&@2(&;EE!CZdJ傂H^T^u%llYKrPXZbw)V=hkDF-:.?T2(cWuW>—g4_r|Z-\?_g&/G Zvp7??Y T7pnZo&vNz')yϧdFKVo.U=+✇uX}v }:a|2[Co ܧ4tjɓϻͦ.(To6;Iu4quhLV?ZivjzEmNNӳrx>ץ*W^vER'<Wiwwllc_=a`iYެeVU/u  ݫƼjANe*h":oUljE4 41N 1x!+g ipOkO'Ph{ 9|z'"f=]9? In\i~dLB+A8ΝEج#Rgxxۢd3t;)x+JQa2mv-0C˴W(^<%d$ ?SȒN iVDMr$Qڠ Bgyf+ vKsuz9ao=HWRLȜAJAHE;ARI$w†cMPZ@QP/~콃z3oqW5y7~M *Iy ψl vJg Fꁘ=V|(jyqM&HƣLvxƮ6ж6*ǽ jK$vXe[b'b4AڒC W0**M8:D+B9i2)z .ǟ$|8|LBjEln#y c" u+1ʄISTg$XQ秠dP^ao:!ye鴗,yewgɠ-4v}0Vrt:ؔP2 Eik*J P6HFKY̔LІ%ecv ᕲ*F QhLqUZ/nͯKU\Ky*`)0cy/'Zl%@4 Et; t>Ռl0Dc 'U@:֣ C,k,sR> d-*DdK82 ű8@[66-D{%AǓH x8u*;Z"4-0u&k^ !}..)zQqI!EY#H_,RLLBvDEZ`uZi&3vq٭kHkFɾv5n] k*)F)F, 5B[%Q h4zK5/QΌvvq08g`Sl)aysN >g;[\WnNpy?zP}@:C.?O!won. LDYEg3%9A |ulPw`V6!\d]mo9+?b|tE2ƱOv-ɖ˖%n%=Vw>U,V &gIST /MH.9JS^@Eʁ)LheTު*YbR61 ruļr}Z uv3MCR>.Nl ~NR.>e, Xuyej 9b) B=9%JNIAI*MRM%b f/>Ȑ5PK2$%RTD=fdžB.+]: d\F{no;l^pLSz_ׅZh.:C#mAc46Y&$1#_}!ИcRhjVcR8/UJ/aNMٟ:Z>\@įuJ~H1 ,dz*h Mwy99]۵Jγnʇqnџ62GVCcyưͧyݐɾ*XúWjғ]zo=a?=6VO|J-$~&g״VY7V>O+*Hw:BTd mpq]z8)HOCWJJ]:e$NTp>.]j4^qd[(貄 E=٪V8L(M_`s_HmzŻ77\ ;aٳM͉_>L?kI5.RFDMҤ4h-f @N*'(2z?ݥKkAnvr<4o'_C _:[N~iՇ KywG21N՗s nS?/d\O=P#"U`8_5GCNJhkgڦ-g)1Q&I:(IuVj'xʄ ;apK05Jv,%Y^9,4 9cΔHPY֤ N8*IUigt\EEGUfJ6Գ<%u1cBx9g D՘\-ufPd1t@ ,pcy嬰[UAϨ5_igA֚ N@J䒅R!%] ɛ(rI1]$3n]j[-SZ1ٍ ^ԌΔJXJ&ꕍ/V*mȈ8^d[9gɶMOvmDf,2{ Gr2/M@J&hNtL=Ŕ$z qԶv,_:PWlc[gaHf)(Lш>t Q$DXr4G;&cݐ+7!jG}NЏR8^o[u=#.ݚ'5: aLo}_zk֏6˲76#櫽m*f_:RvS(BJFL' E1v8({],wizbQ;Q=Y*%Jᙷ+I)贑`;Kl$mPdșM$ ;tDO 6٦w,ɳAњgNJh:$gNp8lmHywd~nt[Kb2L *xpz_ج/:hYv0)ei#I< {LkD>O]YZ3^;`E"]J:B nMU8[tlwu-l6Aqv/lꄻ=3:^1n{l}s+'}wc\뛖?^j:ߝ[ێoۧЮjR-xU58IbB/|2?Q>YI'bc#ycV"ƭjF4nq],ړr7y1>IUMNdեW{j2m_.ۆ~96^^]laUU,iwջo_q=;׫5m?{UL8?U|Cm[uB,u 1-_<Ofa1>\6SknFI|g+bU' ?ZHX.{>9[7`)d2ɐO}HR$)%0P@S̠ciưuM CN]q@ %KTTS:j3j&>/Qh@.]Kwkmx7yߕ[kw?w͹OP(%6#J2ӟ 1鍪ڭϪ6.;8[wZ4`?%E0٪.X5PYwLƑB|q}J'|=~׏m;LFXPщSC[R RybVނhLJ-JJAxTT =иT|T)䄥@p9Yiu@( R8#c;f/6Rf'kj+3ޒ]eyncR8o ]8?^8bcl[af%eB=DTA;$wR\Ab0>=Aj{^.P=@BX0ljBgP.ލh%Èi j7v0X/֗6b4,⊪ZQlE[ ͻL.Zc06sj1ս_Nl!33V4kRdUVH#cb*& b::H3qިnfg)W` "6"bt#ʍ8"}3i^Z'2BYncESR$;&LmـX'urJř"!ԢDE8ZOj}Ʃ8;q:<ԑqqޭkHfd_\Ƹ#.x/VSR[P 㔐V*5:N1aHC;S`xbٱ/x't[2O/~)w|zGpK2x!n~|Gm<яNeYɛ_|B0>Ab*;)vPꂗ #2%9A>Qv^=mI=zaN[g %C.F)P%WLM$4B4 S B,+Myu4(@.ZbLl2-RyPgIXK٘W7g0Y ]Φ|#^__>}cé[@ {VksRu-caH%ŪS.EU]1Ľ]OIWFt%t%Ktjho~/ lc`';D{UVxAĠ9} =FEr)Uh4BkM̨3hQ9b0XJB5978;X`'f> j0fD%T5YfV ;/8Si* DBmv%sJ &sNQ#P*qR vHTBzV Tyb4΍^dUd1tS7BhX%'o@qНif V[ ނ& 3Bq ޚih͗w re|z6 s%_͊I. v17Za2a->^@Gza|:g J`Dlp/wzYJ퓊?{WޘI¥@G^ޏ`)wXݟN-'p?r~{Q+zXj1G%Y\ư.2eQ,RXL:Rx##g >{^8#s…$`6ުZ]V. p?U'Ϻ4%ܛyC)Uٝ*w~[5[T;x2*ہwoʴ0FEW@٫g {/ `2z6_5\dzwz!oI쭼Zm@3 XTHCKK)q չ>V>ر#TèS9u&D$57 b 9a8VX`6{:Kp3"اPLX 1Y,-,xGI4 tȹY8LCMj,erCcp*݌9wW8^+||?޶$ iƔt>IyF9| EB`;XQymdwPPMh|bC-3x`K(v) |ۚc^+,m(B3X: `"2`d ,>VuuN°ˇ>xuѯo?9z&o^ 8q {O\k?m4M5 Mm4kߦ]+i͸f// J&_^%;4E6^&iugTT^u!Y3M~ r WW*PzcUDI^5Sxh6Wl#}ybqz.%g3w*/p)p[.Z|Cfm,zΥK7Ou+} 8CqI_}9r='Bf9rsb#)>~8HFת)  \_#RX5bj"Q{,a-iJ%_N'EjUs.SU+ >E'a$A|my`[My~7|A\NxIWƱ3|;i={j#gKtrIX:d8VtwT9@KG}&lә=[&Z^_+^C_aP0mfd_߭S=ܘ}]i62v5,ZJ(0Ka4hWV$zN4FS6hi0ZJ\ئ)*};2nr)ˣ * $Ʃ4ᑰEo)!xziwlW56'5'Bl+/VI~R;0 CwBɕr]ql JcF1K)`)$Ť=ОjxS )lSRtRE<Kˉ;Hk^M4S oE &1dbMl-_}xHu'K(#_4;îxKn-:z"tDБ>t1L@z9sK!SHJEmI %ȗZrV:'M-HH"kK>}wuu}2FEw|RLYvӡS|b!>\ O+~]1^di4v[ϓ%s re~u4?b<DpgnC6$1Kؿ(N|jX7|.k;U;,7oMc@!ގvQm7.5/s7%dujؘh8ft5P jnR bt Nm-z"HEGQ2mQ1o0Y*B\2oLnfΈjӟ4iS|%&JiSݕs qЮ}rMvJ-n.ȍR[l:YrwΉIc[K(?.қntV憐5]7sPe=[vS^y^[ԼPr|H)7|;4gգ7T\gdaHBDD b FрG!eLDm2W r.9@t}vsdzf#n 7=ZY,8o _e<"k7iӾ6"y\Lk\i4!,cA8B4nC10iNN@+Gb6(qIQ^6_3v!X8<~! ;.?T;)J+ue<* p2(XR*q sL5* N Pxa<ķg-i9y+l ñc\ UGP  %E>pt-i$n\/+g[tq=1W{0 I( QN|2 +RXݮVNڙE2XfMn"Viͮko&pꑸ= &xZuUM{gk% yh,X# XD axVP^Q ڣ6Xegw9\Z$:}9*8˹bkIR6$xA=OSwøi#W6b7ǿp{d:F// : I`ǟrzAC僘muno9a|H (&/?bCKä(YeR(A nS RZ4y#5>~| ?20B h$QEm)8Mydgg߬QT 3.Soc 6J[G$7J!Fk9[t w;ŧ fNsT%ZzVh줟=Q^ V܇j˅''Bl.uv{cJdg'?{[g)dzԭ$OI] ΉuO}AjBΙzi쑗uv3%eZw%v~PKg֋Jm\.9տ<ݒ3w얼f#B F +LJ. k]XpכFYucpp̸K?}tƦG}=cry@9BCw08;09ϝU˳ЙzK-3Jy^DnSmnIOaLʕ8I뒍vݹ\j+%)imoOR$EJJh==nJ/\ uuyn闉+Yoj{^-%c䜼7~L#6sctNϭrN_3 8 [/_ķžgzY6cM;cuAJ~gfJ*ÞIhc{0n||u{3<9lj{Ijw϶0ooQ`~3&nlӆMڝ%v:v_ܗ>ŝy8B$EzoGn;\m#nt* A;,*.r ֖"seW=!Jg)K (/ܮ'xAWC2m7e,ZRAR<^WoIwqqt"C\{]uzVRzir#Uy0V kWMy0hTPD]O!XdB}}q}`ߥ#XJG50QU~9\6e' :` B)u΄<"*$̘ȸk:ol$Ӡ':\TN-S2Ih8=3FvFPu3.JNiʊvU5? [2tPQ&sN d'f&AH"c( bLҐK076w IH -Ѳƌv(+e݃O"0PFhUC&(D'Rze{. x~\LnB%U*$$sF01"3\ ^1V;UkɬQ侴5%=$ d@D%ڤə @GPI` q c=I %0? ǐ']MhDo jć}yNSFbO#Z[ߜUPq r' !OP} y4b{ .DKc㤟&tHu"eFMHhJZkecZ#l zwlyr:oI>`#e TJf_gs33m[Go;T6U&XaNra%dbHMSݓCtUu\U}&Eidd\^{ם7ָCwX{-oW}Վby__z]q{@%&0ZyVx!' Xf.%BPL삏kgsmuܛ<  9t z4ef ޼LӾ2#VGFF:^%*++/“:Yl.`3.V=+wIaiИJd t\"!sqU e6:pQ;iˇ4{|UҧQS,~NGt]1dLikG]`nfn~=xz?6R6/mIVuUwuUSiuݨ+ԛL _]݉`![]ݍ\0Fc9w`ԕ2u깖RRW`kF]*"]]* +-*~@]u \*{Pz ]"QW\kE]DQ T+N>㳺ɣs$g7j/ZuvG^&\Rh]T2+oETboBhВ=:# z]s^âCTltRlI&7](z=*czk^Q1w=Lp1P nVBdt" מ3`$618$3ɭV77< V)8ƨrʉ„Y1`g)w权%(:vHб@ 91r}FaIZ[ !, 2MJ(0gc9댜-嬖o߁B70*K{`%I-=dVs/M% i̤N#$"  "Х\~/*-#, "#BX'όLOw*6jux?Fܧ 5^$X e 9z2 mF2 P^ZC#9:  8:Qj>O|z1E]-B6f'a ҈ApGMKX,p-d]DhyE:c]Ou,|[ Gg0ddLE ؘY]AaQ*%Uacƪpk_3U:tҡbbZ kp#PhFtIU_R\9.i5(u{f CU;VCšA%'ePrP F=DƇԧJu߮:M':ʜ * lBsU(̹kTɜۉ'9V;dhV!gϙЂd%GUΘcFHKihxTLBZMGmyEڝ-M̵аvo5LJ1A=PrY-d@wĻՙȂc<~ɜgh1A}6c錜 H=U٨1K$* A Ld\*)EMCH,"QsneBe|]ܒ4 MYr‘4r:1!H~Io$&8 zRu"*@ RS:cJPw!<3#<:9 qo{JC|Ҵqt^}VtP*:*bJr !ywռi:N!qJ!>lbܑFAomTD9J58ui}9OzSi.5ͼnlHH{CZKrhu9OJY"ޮ S \8~KVllf˃ z7 `C ?ZYV:Gѥ:l"s F&b1 1NEcFXn FY:k92#p쐻Gs{tx{mAwn^~Vj~nb?n$ ђr4AҧRIrwˀa%, ʠ#Vf .7k<J7Lyi4.iRT~0U0*tWyQL{G;x5jCy5`^M)JR (lTp6 98rfO>wRr ]KRzʘ 9QhBk# r3rnFZOiǔtWSϷ`8vuJC7pA$|@g#AŀCŘN5T.6׍pzR5w~r(S4j(T#S)x;}X߶hG;XJsȴ,fFc [L(-˪S`Sr))aƛR Odyo9b"3޵q$/UOK{/E-[d \uC)}%u_3F2Q1qq!в"BHH H)C.JuԶtb܍R?/Xl}슈2!bV(.:*k.jpl4"oC%U" ;Z; PBG)85Iiʃ,eqyNyQaBsd׬|9#iĩvӪe}u[%"Epō*=I1B$NVDjFb BIn=.@ȉ ( ŽK]C-rr:gUZ{ 9'k\7q 'A#~ѱ Z\{l<IxR "aAYWq]%c5*"Idtvn r' XcR Rs%69ɝIR" P#V'ӾCI'D.g $L0ǃRL DL ʢ@k`*QfE bL\٬"9p֨h(UXj#ނ5Ex)Z5c+n'Ps7*g}|<gi]Ż/KV(o) s^iY .C/撻L/;Ve*./SpEjÓt6|Vo%Vz9+`ґȸ1LVFTh'e=xxY`?,^P ҁ"RP$Ȝ2Ii] Aÿ$D@e:(:xd rL *@7u` ' q1(J$"n}}ҐwtΛؕ\=%]}]D]a6};֭6kW&ߪө׏^+FP..ߏ{"|P@h* O .$U%uxRP;V}vLɷgy={? 3W,o 3Z k7>8WX,!4ZJILc87}w p6n5m o܉'?]ľf{N>;2 jO+R%BQh<I2-u4'K>{Znq 0<0 F9AN^[Xbs5 k(·U\$e *#­a@b8e9F:HL#I+97&OOf!׆Pp/Vt\_뗆0Ub n+0tQ$n`O T3nV/i}}0.E$)J^ZdLĬ8i,54BR`a|"JNeNX;āfNg7_u~0.w;Ë)ȫ}J2O\Փ?6NF^Oot0HeQ|؆Ў|]ەχ㋉I+LӛÁ?ypvzp{7__!E9mR\#vt_`xV!0ԭ KLѓ8'voN޴nh.Oj'h[~>){z Ev h9/6x|kn޿՗4X__y,]C?]yROAK  "NpzkҋP'jl}͏Aؒ!y^ g=;嚗 1헎JBYآQo2oj%|xM{Q_Gec1nNLW 6S1Kd~+DѴ Wqoҹʵ4g~˟ڶrTm-ny[.Che&qhQm;enRd4peSUV{bLuC`+5K\)ږ]K{D׾7=y;~ck}&vzx|#R)e6n=KoVS,])W.p>u|iRnU1WJ];Mj|lƑ j 3Im^yʸ5Ԗ{yߑnQ 1o[c62&lVrCCsmNYɊ(a,2Zdڑ(L?ۼlŦ;7(`|_g9 X`z0*1L UR,+E =qX#oxzԜY#Fǩ*hi$ґ`)%A$s,zytRզ&ؗ+Q ʅ{U_pʱkR~iO")k.DkFĪٚUSWG^9t%4WBW$211L~G V{FJss>Z~'Q53ҙI ʛELSBC[ǂ%(J,S6S-I&G<\3tؠ6\' b zNG@(H3Ֆ"g>O_6vV,ǞL5#Ke۾M9{\tUo\b9!6P.WځD).Ӟ'"I vY˃2XFF"y.%Mm(Y܁Y8QȤ( MPblzS'N8@`Q0/LNjfu2ONs\r]aX(lb++Ts]')+R>+ O! Nut" :Y /@q34N/*wzaTF{I1o}N(F\ CKaFes{g*_MlPnB$l$o:!뤭l` ́d{2A6ϭl}娕 lb:EL`d{nfr;Md% gӔ$" 2imҲ)%sX9,I[!Dg(gXi'=wΚiRM6T`L'lH6k F)\H;ʙMхVۖu=$gE%O^hy:iŠFHo@g(h2sO1'qkZ:%vff'Б9H6(t>fO90z!uǓRpv1k?'ypp6=g\->U/mڛfGx\(fz'>WLoz٠>qiP2G5 FIo~=<ৗ?_?S?|8q~!A$y|h}Ӕjۛ7AlӴ(viWvoln-j/F9Z{l3]fhٚiG_AϚ Ӟo*TWFc6.>džjǩצSn`#ș(ϟꝳ_^>ߪԻxyş""pWL +/zoVa ŗpq$T NS Ӹh%|RP ]bt^@n4FQ,qd%ҵRa\@5;kJKV5ۚ|N-.fq6p+&ޚPe><&Yt޶\.Ql9םYS.Wy<r1>5zfϏ7v]:~ Ǫ#YaV43y| 3UJ7~jٍPO).d::2!GYJiCvH|h*Q $+2gd''H:H#.Os}ga2Pc?~i l?Ϳ/Nn/衜5O˫ ^G'Qֱkc۪_}!Ͱ#fJs=wv2{n8 {,YsteyVe%wƠ] (*+EF1ɕ߉IӥT>Η !JvKނmp{\^.E}y^ip=kz(Wo-B/7lgfkIlv>+lkܲLm?>gS>n y>+ 0=v@Κ{L[F׭Oĭ>>x介H!֕O|}>Ȼzs3 tGΡ*;?7Ms]0ݞlO49Vtvffb`Fm9/k9(KyKY̬ Ue]9U`Yɡ g .!rzޭ +ݕ)5`e.U^xeǦ2h*Ǹ ȪwkVnY}Lw+:E&{?Uƫ^a?zz(\ƫC23 :@[\kˇG2w;-3.q&ؼ&+8Ӯ=*ut) `=/s ¥R|X? Np\i^ ^֠YT ʈ]9\9,#fCb/ cPؑ #mt'"}N<߉7VPۙwlu3-W3:y2;In~*f=&^?׃{r pD̠*u湇Lm<#yXS8jxЕfdmN 7s4W? qW^}׋Ά A{nҗL@:(+zxJ|0 BY(.˽-hǏ$y :3Scv^YA;0085ibقl&ZlT LO[Gr#݉7uj0VWl鯊0rP6ٟNA5̮u< 4_{>RPrl;vrU V}KƱ:q!y5[~8ƿVVkFāmn߿9Aqˤ!t|Qe4"5*C 8rv\ &S;Jzrj!Ϊkfv{ԽZGqA<sJSvvSH($ⰦK&$BrN%$j9>$B*t*ݷq;V;e @qIv`\u Gp50 :ԌŅ+WnվUϙ}*W$ئ+_+R+L" &\Wr۝Uo Ԏ5׼}:#c+E0\o'V~;ls?\بgE{iPEWQ˪ @Wv .+~+Ƈq^o{,_d˛B.)/獼olv7[)6[/~bқxn}xjgﱧ,755U+`<2nhWQln2g(*_ HT>`{<-0Ά@د{<36Gg*s-tJ|8B_bAیBi7[X,z_ge(^JWzR([\[Yx t(: %R\J[uν|ZC[/7+ICZEY7$m 5_dWEy-9cWpP* ucۿ"Ie`0(e0ŀ$#Xr\.J=vcqIJtH(T|q:C7v(ե9Y%$Sr!U51:*Wfq!R WkTgX;&[;c1HկOBCsZW$j@o]v*c̙0!f-DZ˭!BMTDZX%[*c1&j`5m:h1ѦdSo\_RRq-Ujk-`RJiMreKA!ݱO[^ ibbb\rV>cpODkg}okufȺM(Kc(Ʒt cE4P&]JR(do=pB#͵:21{ oTl}G!yD! qx{f]=hWEU 16,Fb JPb, Ѡr>9 AbT+s9:Z g]ڪUE']Qcpp߬#).Y9i19Fd[rcF46o3sh3!%*_U -%$ QR$DHb*ڌKA_j IU|^E"[fIՐf!Օo*dlnCM27gYB Q|`˸YǢM\nB B!{ ݥѷ*L#mT j'B\,ޢ@X^ XTj`t[4sXP3 %eâaD)SiՁ!V/ěEeCX͝@kFh#sDQE`\ZB64GV5YjH7e"pօ l8 VlztZ{)j|"rEFpB5 E9Q {6s`q]Ò`; u XU!\ "U*VnN"22)!=C&44:[ x&$X+ԣ5Yw[4 ,ӫXm d4YFhe &ڕ_t%fP57ظTG( 2[Q! `-Av%NwȱQ |Mn-dD5R rљlڃbUM52d*8(P.hq`j YGTB=@w/:=W  i0ԕ/h!8[j{QR(X}ݸ<ݸt0+-ƻKV5hΌ;:kA e 1P@# ڰI&=CBck!z_4y@gF\' j_g .UUiuJhNo(ƌŔZP<Ik!ԄK.m&l45|r?h#+WF#󃳝w mzLZ$,>&eQQ XYǮd6Q t}1[4CU&@1$h6~χA""$7]uGS}c6FD6=R\bQmF'-R @5b5R0 Q{X1z. -3-hY+} By@ 1^A";΁AbrrFVVZz!̓rFHQDƦ xd-+_ca,A,Z$@fj&X9eP?= VQ#;䍇""Lk;By( 讲|,4⬐ -bl''8UKB]Ŀ?(z.:68MV( PfAڠRԵUPJςLmU٤E5 ~+FG `j|0jn6vٗy`;Gb м9{T|1>ݜ/۱LHjQ@]wP7םkT3 (]ۿΊCŪcԮaZ3YkI.rO)Vk+p6狋qV^ׯW!=?ym5f+nn^&{'Pz+N/ m"-N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'ЗBi@qnPO 1}N &N/ =8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@_GITq|z8j:F;^N7V7s+^__na[)`zB_ablWsq (җ`\Gǥޞ_ujMe~oף]"e}ڷUWSe@ioPgy3Ռu9m޽.٣F<

742uӠԬS &@xqbQlOz,5ʎO"Bj֖rSʵOe Qxʒ2+{\eSS'bJB:,2%3Jm욑߯ne{4an2ݽ:r :n:nߝ iEz1i;)I}z )-M~fVTigM֥mh g{ΰx]wP/ojg? Wr8VG:||,;Ru }?.+i}Em!r]>2Ζ6չU6P.6`&W<=y 7は:?9Oh.=GL$>-Fӟ.~W֣SԚ}YEΑ>_x:g X?ղFI xu\M{Йd!q޳q$']c>Nӈ \ ݶ7y\#Jh9ͫ{É^?{׶FdE;h㒸l>otlئ(5II페D(ER4(n BYU9\_Lq}%9赵Y+eXlG:nK(zoTK!s!;9H~Ych^i붪r̽ݾc=;׿tHn )yJa#Rodyyq;vr◝ [[ͼ;;. ຎPVa] \բN\Շp ݅</q-]Nsg#^tuɮ_bIVyO;B tqt;+H}=ճ~փ8/P}X'lFlF݅'}\|zz֝% )m QY}:}zW{6ӆl[|l[=cVf;"7BZVy ͑UH%lu-t6!,?ƞ\7:%Km.\ WQIs%Ux:sP>sP>D#< " #SPY]WɊ2kA6W*`Ma-:šsMp(t^.銶V= b4m 'ڔZx5RW %IXσ> 8S$F-[1$@-ǥpT& PМ/X/WwAC#z#z v)$0Ӌ{IiTV)dƔ&sGD5FbW[H,/s4:vgv8d!@qMi4)9 2%j]6-@] 6@cc[m$T2-#ZsV:$kHˤ=IPAU_n/ :NZ N_z-`jqh#UQ:jQsiQ>AT!gSjVt)bϵבc-~[N Q-c]5ig-z4Գfg^Ԭ[ן-Tx0`H8*z@ %k }iP&X(dN3"\%5@K(moEc -z<9δ#F`I"JecNl00>j5Zi9yȓI$Dk}1K%j TrԡLiG r^~5O a|6-C0WQ.͐d-^U5?׾*3<sls=P9ULcD0=> n\gT'}.2'",8&'MiTvX:hBNX;]ԍ58히n[sr16mb|mn'$j=a!b=h d]~s❿_6D!,d]; c&i牵0XŇkpbmN,}XF=+l6\Byq[$Zl 1q1 os1)j^^LSk!︋ /BE+F/PsU%Äh8Gf65r{%6|>R5YA#/{#d*);K/DR|oC&nC^rPU HJUulJj9~1lNrpiur"9F%-ȩT'rzgJi}bUwZ..҇j{};G8d\r(֚] Y)E PRy>:I,Bjw!aj\.JH2=jg2̮F!4BFel&asԵZgHRAzmۊ{Wv9].oj tj3-CO lV2(]2b ZH~N-.dh( md0W1#.2d Vl q;|);xn&agybj7[ZƇҌV{DV\'%'E,JՊ$RM![N(ahERt!2@,:)dk2d$J GQFjs3qÎZIjbS-bgiG8Z}+*^iGKZM>@! Qn; E]r0d ;Lm ΍ƨA _*[ڙcrP~UR",R VZwh&.u$`k*9.vэvq{žb4hk+0TV!Lz˵=PΌvvq088Пf`J}],#1Nva]ApcE?ZZ9gW3v5/|nlBd*ȠDɂqQFg1e1A) R ; $[rL DUjiCZŀE b (REf mR:d19"/Q}@1'!v|r^*^Nuw2 Mn{mP*gp^sd0(k nhrt:p) d)\0g %ضWi"KQB 8:|(jr a:HJPsXweKg{c3Sp~1K̲3dn?OOݳIo(=JY>1H3Gy𵷪H!D!)tn2sQ{8lxֵs:ĒTfgɣ+ GG4p6p5Ȼ&tN> |D֢TDi}@HTƓp*fZs .˫ 0ܺg>OuRZ6"׆dƶf?} `V]jÄW:a}ƴ;ʧﲻ EƝ9n E%ۖ^y;@)諟| Tr]BdGQ{62BJ BXl 6ZX(ʤbYdɽ'ZUSF FsR!H8y^]OD!u8?wD}3B~SUDntx/2(M?6 QA1jZeqWy태Э+*ד=A#6KevǷ#y#~zMX  l{1e,LɌt/[t:@I#!J MƓrrvWs¤YՆdH`EH%ei  1YDX4iJGQo{Up!+u dZ>EurH> 2mU_Q[V8ӒHo/l,If[PGr~3m|5H5"Q2P<3o5$%=It(]OnZ//~Y%PJ63 r8 ᢾ3Oewy4~G Osg#^j./qvqu{'󟱫oM}ymߣ_=̮̾g=gU:ہ  | U~i6*컘hqcÂڊ^=Xjk}n#z#HebWMN&rc`!-@ eP;H7Ћ Ћ)VH|4$ÁhV~HZ"<*/92\ mccG쉅 Ql[4/)vr-Ǽco1J] lEg3<:pfP[KEXGp,NDYPL${52$ SFm$$Uv9IFǨNOEGdy4D=TF`UZ9 Wv `wWGVv?sR? ~˪duv讐|}3ME=}ZfKZSݜ)B;}K~%t5v}5˰\g,zj'@ԑ| V~7bL6RtBvTF(q5'B6:0O*YWFPv.W]l4k"(HiCiI}<#nWQȽ-j-e݄JkO]gHfJel84hzŁI=L'q&Fn?̏O$;b歑C-?u5T;'깞뺔ϲ쪻S7C@ֶh{6-&>:?i{+Oֽy6NE>2q3$e7~Vayy1G/$}SPR!78&S q-0!eaH|Pc3 `Q(/O8Nq~xW~=8ñt@HS^V{ǝQZ2.j +kFh[qqlnmJC ѧyT71V)q 7_&]MZZ}R@~_^G ๏sG$>W5V[^9v_E8EkyG070 @i >˂1"4OBgʛڸW⟗dh*j5 j+XH)\SǨV]?S~f]d-U: }{1EuKv]m'YT \tpracD`A"pBd;n@JtbP-׉m֌rxN;usn.>C87NJ뾓hF^&g{#7G{lOwW3khpFAELԏ:oM]EM,/mj$A:?`Ku=G#6Q>yk$IH;,zs9t 4ܧ usp2y~sɧ7!oT44_ĭt?U-V_LV2TZ;%o074,0 R~WZ+SGiw3ŞPHg={Z:e%w'(>ɪdӴWl__|3HWiL"kZ'P@TV/91e)j'8*c*A QY7gpم 2ʨR6w *B?`&*K۵-$ y!KLOU[KW1Z̺loS;Iy}l>,= ĝdJiu8gwt^)ল3\SZk7=v/{$KX1f E<_ ?93n]6`<]*?.]_qT+]k;_{W/=0M[P)]osxw!vv?Whp]$UAȵ%R ӗ=~ˤ L"s\8yŽN9>QƢȬZ2w<,v Q(p' iw(SS9՚HHB[jZJ- іy&sX &g<RVFJР$`)RxiEk1B8JsBBd9Ժ?zix o~Uu6ZQZv#H?lD5[U NpMPJnڄ O39=1^4&Zi% 4ȜI*r͙A&iigc2bF1,(o)hXU`G0pg:*.4 eyHcXDJy9S+ !$̽6^Ԗi%KרN?֥C'Cnuy*MC02'O[ۦ|{k,Si pQQ؟*')t{n>_@޴?vR^p!PrR}O\.etQ &?iVJ^l|g.(7?wIZ@r'7ȹ~xG1wXKAqdU0KrO/|]rp^9LpzZ.βa1WDa$Wgm+`ۺn͋Ά)s b9\OF=7ΤN"hN^*=;XEs{SV9L$#+`ә>'Z-nb,bJMkŌe[ [SyBr %h3 ]2J"άHi9eT]v.޲IU1bk[hX9ɔGNȆ H TZHX뢷\Q"c/EVE,ֶ&֌Xnw&v h"[:<ӏڝ~f@Џ>gvD<n,lXNx@#Amԇql(-MwTMr"oZFGmW0u+_C~..MaTtm ݝ/йOOZEHke-e` yM>jbs!A*1j 3ꄷXǷ#I×󖯠Wi TRdf .aFi"L4 iy 9fA^erx`` KH2vO`·WLgNlOG)Y0lF8Go^nD-F&fT:EAH@\Y g*U0;=٭W`YtNpZݺȷ/<< צpP ޖ_O'fVS]9]u&5]?* doT67K n &`5fWiA πì:} ^K^4+~Yaz)̼U͆rr-c13Iՠ10 f1O)s˭#Ă`beP"EYCi[o.udoR&&VQRixƌƁZcĜEfYI/n-3-S6Tt}^} e YYg_>~6O`݇賗eO ^7,rn?_KH˚7W y0qr۹GgpFZ؂%dzxNt%=g+>/p·I+jg9V\2s.ONoF c gVhOonS۝G-#j#Y>6 `DtqJP;yd+k ;Aޠ3]=19u̞p _hvcK+m\;n})&5K98D(e5+kOLIiy'9X+17mƗ\Wټ}DϨU[Y炣bBw-4+E j"uYU\1xVD-zj֪ #$I 55Bl;\%*kǔ !*W@rv*9nWkW`MW@qMrpTgWP(a&9p]%j뮀J:+ճ+E83>0+Ihr^sds¬0Ϳ8M2;ve͆qyTx,=t=\0}z~Kd`:Dc`ȥ11$jsJ[~0)+ X7ruvʹ5UJ8bQ-J=1\݋\p?jYDG%2\֝zLM+ Xq)r40CmD%-\=C"G`,pc*PS %o;\%*j%n\sݏ\śW@-!l*QI W5I GDPM+"pܶ-\= \q Q + bJ2J*pT9\ +!4]%rh \%j5v*1-\=CHFW)i0eDTWJݪڟ#\),+Q1|&IrЌ e%=2{PAGf/٠f]s]7+X bs.`9~ 85qu1HY$&>pVz^聜OS+|b0tUltw l{qDǓ]Ԫ;E!;Ky$"kY"MLBBәV0#)6gF\+Dcrڙ`fש3UJl4T,T#{2ހk B2Df<HI['npP,fa-ô]y5`_"d ƪL~vim.h9F Yiԟ;^ v2Qlݨ[0-KcmHW![_d`s9/f (>RݯjHQc4]ĒwUnOEdsT5w/{ՏQŽlLEa/Ki?TQVv܌X^h6@ImUz^V *:Q˧)agv5wn)'҂H7ZJc3߹~'i\OZuhVymzla{5M_Y[q~mO+.[WΆum!kEz_:ocO1pvaOMw+tٍC.U(;jL6qzC*CxSoO]s52ydX uŎX!M`u$X -*TgOk?'Y}A;c#|ўsi uh_%0aOJ BE13yǟB{TʅG*_WGxWp}^p̯?hCk޷4[O6Q T6FfF&;I#+Yx-cyqVWen8K *HSgUzG_n 1*~Us1`5@㨗H9NH׍+HdPFb{lOH"m$WFbʪeZKy9-͖̉ƨ,MuuEGZף:T $~~|n57`nvK&<,2fgU?5RX=c8AO7@o@dm~=8Ѻ߇r>3ѿo=hko:Zߜe&-ucX>ʹD%.qb)BM`ւpjǫN86Hs^ R{\ W >my!hNIp㹝"oII4(" jzi?%X馯kx1N-4tbB j_KViFBN\\#+T f\W*+ \\J%EHW2{b3΂+ӆHp0lprAX."M8Si2ڈ+0wplpEr5pmje0NW-ĕ)NA\\JWV']J+3Z+FWU HTqB\yᵕpi5l+R GW2@Uq@Urp$vE 9>TOKMgv^XrP|VDIcߌ6,%Jyr0kt593J=Cq'^RDCRTr;\}M/pe+T"㪅RV/ IW$\ Ղ Tڌ6J+N3 Hc+Th +eZW6q*jg+K'\YH>"&d\W #\>>"Hb+Rk]W`tPb2HrDW6$2*sWmĕ`+[\M4{g޿^NX!\I`{XQˤN~LoF{e ? ͉KpDe|e3dNUb(7:4OM/ϾyoowzY}Z=<Qʥ{O杻wZ\T-{ehk+J? BB )zB[i4XHWD NZ ^sZO-ʼ鸕S ֙Cqx6 V WVC"y Uix</TÛچnTBb  W^`'%W$ײ!u\J'EU q6J2 n 6Jzע'YrW<(N+m/qjO~\EiniQWg31 ؞H0(6"sަ(Hee> =B ˦UBpm(7Pnȸ:pVsPp`44\pj @BdU qe2 >c *ԚqE*d\WNp] Z:H%  W(X*W$Wj'6+R 6㪅:E@,TNT~x\./\N1AjTRvŕ@r.Fg΍_.8#[$-Μ=OٜigZ[`3&f-u"lR!bNj^(IiFA:Q4?*F3(W;6 7gBQ 7/ӲeO(oD󛎣-U uøR]CQ*I쌄̸1 6:Wֻq*ȸj#p, @e+ Hm"㪅tH2  Pl+R뒏PeHm!,2xi mZW6$]c4W-ĕ5zġB(6"g=Hk TqB\94wlpEr |tE*}ڈ+N4(N|VI\pEj]y*xvB% Qn.EK0TO||rĹIg]X#eX.\,rS-EQ:>I$U_O%[p_-2c T(ya܅=kFqV>QpZ@FY!.>RdS 煛3k^wZ6O*Nkz_\hx9Nmh(_JθRP֒ Hn'PԂKW2J  ;F Ps+RZƃJKy]f O5 H TzqB\iǧ+,5 HmAR 㪅I P HIWrtF\9%m4w6B ]Zl+Rl"Af\W^P|H\pEjNW ĕͯJ^XsA[ofn{f(>Hb"Jjm`R y{izJ 8JMSko2|R&W&j_KUpm8ZWֻq*W-ĕ2c+\\/\ZcRLmE4,ƞ|Qq1lpEr7J6u\J2Z+cQP9 W$W\ڦũy0F\YkpAy6"MS{ Uzۈ+ {W$+.B&?$Ae\WPHp3wrd]Z+S6K0`"SRB$< !}%U'%yDns EYWݲϷ"oƢDSeY&hp_IB z^h鼐D ߦ*xwKp3bcY~)hP%[,5ӝʅ6luWF4+ o@~P?+ey&Gt(MeTW.*뢳Kϻ V7],+\_\?tU_I{caE͟i#@_ >UEnR;48`i5y) lZ%J*uޤy`w?5^, ^߯@g?>QZ'3/ i/ k |-Iu)/>G[O l7y^\~_OR>I#/ǑdUեu{8Ҳ7Q*q#S('"YeR_T lM۞ Z̗sϣY1RvvTy9.>ɂl"7^1 e8_yqhv?B| A[3|J sp7O(Nphq\#mQڦv?ũtF6"4949qЈTTb6L00Z-Pm] :|x"si aK_e JXWyP6H0 'wU<?6(fi em{fwr.Ao^qn<*0<_?y[+7xĉW_|%3l4e?4;jòYޏx6F?U.JuwgaV; w@AЕ-S~YVЎ0RWV~t8X6DZ8 GdQ ,:T#[7}g ӛ%,不 tͮ_߮/Tߦ; !VpY_}b^~?0?t(nOAPC(ۂ6'b00b.׏dЇnm~r~|G:|}C "VfNn‡Au/n>Vs曗p-܎zz>O!v#ix n6՜IOuY'֘JeO'ʐf(OmR{c呈9$=2WVFsR]C5ny5"XymdOrrخt5NRfz*O&!}$.ݘ KgNjt'%(g2 l'e̱xT j#Z(egzƑ_i$6yMz9"(ۉN'|됦L\**օr Q"`Yj(92p1C0&)k%l f|d/_W}ʪUeuf* y.*_9$$T$))H4ː L%dysDɘPQ{ٽnKxi9.E.irHsc  Q=< ɥLi&x**% ɱ_uwpZ1AO3>I`i'!sLiJ$b~lp'v]%0d$,WaDF=w>r^M2>z&NƄjGcg$@(>W7q8A.G:ì_ =<2ݘX82͞jc.@4&dAX{8k 7)ko@K/朵U!f*?yVinus ZOϳy'h*I`afO*;jz?'f}~k^]12~cm\bV>,ZIk"ΑOd&w뗌 #- J%P< .To (O+1UjvU㋝BTP#T.ۙ5 xk/~zo:g. ۚl6xv-wzyѸ"K󇖮- tb :mgwcP+>WA"f'e_M4I(^{s.⬚DqAHH yd8J砛X &hp0 )'3Dp0d1H1ÄIy4FqRӇvv[EY- Rp Xd A)\BCQ"SXnry k~`ꂟ8}?0`gؼex3K7U_S]YVXk2- Z^T?{X:/$ñ!H/++eHWe#_:1Kpシ+kt˱j jf)/($ \ᴬTEBTa rK*ySJ i=[%cK ZB2?|p,y}J b{J`c رI%I%JV"CX$.DUL8eJļ\d-ӕS 8l@k?yWZ x=HFFl;ۇZLW-+:hBv ?wKCÞb1/H7Ic%]W$(EД"arc,cYSXwwlrAкJ)D1$uRSzݷMcqL=yIK`<{1ٯZ˚g%e$d—n8%jвu%oI Q10h/6NՀ&J6xe\` cqQAeeLb I~U4$YkUtYܾveu$2ՑA;7$u3ȡȗ &fA0+qSS/śT/\zw_ǿ<7OzwQΦbrpp!/q-Sǁ M1[^ӼͷkckBӿy*6]K5d%wЗDp;MD}O|Z: {eK$!~x1j:+V\ MSZ#!Y2N"é>}2i j>vXhCa_p@)oD3}!BHK{$ly>cf&ٙ%TRwж}L |u5ǡuwz1#HYlmg1{M5 52ĬɖYmI H$jM7d1R6tc10YrBJ*e2@>65vɜn^{WovbWՠN ̋ Dh'L4Ttu[Bq>= k*a۷J(fe%!rV% hGzK+p+!àMšۋco$r ^J1 8`y<̧_&i.%{9'm"<Gkhl| We&Y<)o3{$Ƹ#LN@A3^tK7ok?*8H}&t#EנٜZOΤt&S:už+әΡ;&̈Y__X!ͅDM(md75dY[vŋkW Q8ZZ|o7~r6ط Z)|Ckc6G,mҀy@jh;ci~Boc4wy+?}o4Ոn>: yلbDmO1h؋&jͫ5^*ͽhgm?MbTҶ1Q, c U6ݻ'=Oʲ39MȗMx)F$UfxEvuS$wMumiѥ&iLNjVG/@]?m.8>|R&>,I@ù'n\kGHsIx2~4u}Ojk9!*QrPuiAMooըUM׵\=ҸQit/t|[;{X~ \dxtqMK?%2  'sT#9W8eܗ&~`bHmEUЫJ[8e13Pߓަw\UrJU*ԫzU hRS{6ѡ&3Y'ypZ8r9Hgo*0B!(1ky/--XvnгD Gτ ~[R~i .MN S:4&Q JyY=~v+-r;[q7K_O3U$$!@DH*{}jbbᔖBS; \zhC'7lƹ˱>[UYXR`BsDR&h r,yR{Z^tCT}.t4[3ۋÂ'i"HPFH˼H3Ul~䠄:92چn}M#s@{}`%W g+2&{:[OA HU;[:F5GVo9"N׳@v!*(~;ր>Zo$%q|Qio9#q^>6]/ 5|r7p΢l0qU-eOl! !HwiM3Z]Ui֥ūf6G(#]4ȯ镸,&'u>d"nȝ;c>VT 3Gǜ 4_6z 1H!]IS%]lҝWלPه 4j1 )4er0]C=9z<ұ/vdMp0nnv OO֐%oy_At1'xRa۷yC:6tcJ8e+YX%iK8No=fxfs̫00SG>2eF9z!Dȁl[ rX]c IڭWM;J9;Okm$ا y= yC ,UںYRnaɖUUK$Kn (!s9,H拶ˆ6*CmX{4{,H:Wk5 "Kަ>DvM| 3`Y}#ǐ57%7_Cu{=kuVOASV ^5T qtuZv.v؁g3ZӇ'3RÅg;+݉l.5MYwyFZPd~oAeM]*U-9J_EjgUr6p4%:LF+HDugSSϜz#:}"9I3<3X?S\aaFISXsQƎ8oΰP(TDrt6Yz d߻< ,u]?Bd`OC$H?G]O*oD3?hd8}Ɨq-?Vّ\݁Vrk A'Z 5Ţ6vHn+V ?{02!bak&2γѴo:kFZbiDJ#I:xMu [;y'JVJqdQFViY֪TxQ)v@1-" ~T)O͢)> ){P^vj ʠUd6ah' !y+&>vA2z~nr>P)1"_m]`٭ʣwUZ+OBiT!Dm29Z-9ỹfH#.'scBr10 0uޮgQҖJEw߿,xвsu{Mi9((%(7!%T눶ΞTfk=߇ #Y~QpA0yjQ'!czxa0wqN.ɁRCt &,w?/d9zd᷻K}'UH<1|JkK?<%2 vSE$O΃C/dG\3vF%OJ.tK4PgkIvtJŵ i)h&a3tҽݳ1~ɽ:"$p5ZP;TkgKt¤_iA4Mb?5$sWK煟03{;!;ʲ TSC"H(UynDѧ4kS#25cYxGToDbZ&h{ݽ7'SEZ3eAжىmca S4#fx=64$Eק]6ե}Ǔ6GBYpH4} &RGv#,XDz KY"yhM@}2HqF#)bg؍5J{ʠPƴAE5Ń^'x7Ra TS Uow8r* r0 DǕ $l5;Dnq b4%y3$w$ʕzsv ?pHh>g¤q6KzzT` `ƂQ0≠>ރU#|,|5zmYuvlG#Z OtuoB9&b=q(PQiRQz4 I 9=RYĞm죾yրK=WErݿX);օI] ￾sʿ-Pcj"82yN8AV)bR1ŋ䍄7G5J ӤaVZ\ *)*#sx &`ʔhF}5Ecp b}zUKcn~{/d?:6rX%%)EHx,n;@T<33.Ii:3Gm>XXf /` #Cֹ N;mO36}Se_wהw~r 2b"C k}dPKw1Guja"j7/7UvR j6b6=l4X] e)1:)rRϿz32)HQ^ ZEh_n~&_nrW[Y+ۼn٪)o]{^^][}s&{UپeZL^-E_:6,ٯnn9n6e`|/oNIu*]9sǿ}t~Iv;lj"5FSFOE0 y?DppN*2}qXp(Urό/dJ0{ɟn YIՇxAԝXj Dޠ+~xzSUQN@QTIMEڱ_^f&O;MZWoMsEMM,J.-`QqRR@bc*^5.A{ AKTkL0JJt( &("ePIVe9ϡ2%׈Wn!աnwPe]95"n&I(Yd%J@2ƻɷmCq;«`Aۗx쏭uMh^C-s7012RhJZmw] =aZ, T&'#'2cdFP-/$Vn[ myt( PP=ߒ-_/j`ѰnY7^϶|1 Qw&H(\fu]ngM>sm> o6J%P ̱ JA G k*ȇl{+`gd5KP?eZHf -s֒[bPJn 1>ը{qV=R&ߜXdQ:c+[d8=Ok\ԝZnmlI,-X˪fGzp>[8QRGV)jG2=ɷ\!Y]dYe. /ri{Ni&\cC%@ObN:<%)Pw@ҩ[Z;*[ ͅqko'-s|v.C烎}`MJ "1nKڀv]U[S RU˪iv(t{d0a3F[KlupASNĝArKq;2M3cȖ6!(Jѣ;Ǽ~T(WVQ&rtPxeEwȖ =uEuޤsGOt uV`䭶]:7P?߾q/hmzovvz߽$DqKo^fN(0"9㙢c%ԔX CZXv7X31^:t" h닟NoL݇=?q OMFꊅ8g݋'2Cj:?V`H`~4tfZ9.fߧ2=|E鑦=ee6fy}~۬JV T 9%XqKrW- 1𡑰S"ψE{3rŵ?zH ~2yx-7|6tKG+" Ƃ T񂲨hEv>m6' 7|ϛTi)I=ܟ˰AࡀE":bPj.81˹DVLSTy.15r8jd5;,3xl::|B>8vPd:aD4"uTp᭳@]Cjth̃AhqvFx!lu 3݂ ~^Jto|rA >yG汬/?#$%rSUY,2F2+(,W G{h?9j&{`ۮ C{쌲Cqu;f]S}.r Ҕi9&meVClОҘ#jE5|Fny=>έϘAx*}Gꦓt)OwEFV{&ŔlUH=85exhTPѵ}WP̡WºkYTt:Q5JJa?bґ?7o—6 ;9yGԎC;ږ1~|έv̙R΋$%*DtR miP6G>#hQonB^fgZǑ"4peBR|n{ݧ;@$RWlK6eY-;3A" B\<=@bKJ #8Z],{UPW·nUFbω\Qꐴ (*4$XjbKqkp #a4f:UXekU5-S$J&b9O ǚ`kdђǍȥHjH.;%}m5R4PX`>a魓hʥWqu2]bҭ5q|q=B ńnRR":Rm/ǂ˰沈ƀ˃[WL z,i%v%4;]F\h:ec^N<޳8uWg吰(2yF!2y+Z6v6:KEG2rj8R`!JxX^dLJ̿Ё΄gvo-؇G~D}Lqεky%t~.RW`!}Nh L2!X*ӞDESdHp6Kk'˵aV>}Z4nm-1X]hx.X~zU^"BBӛ=^<.^r-1vؒ@PqI.K=ygl ة,c+]?F #G#-őCL9lM3c \wgfn)"WC&g|u5>e<Ϋ|B-zҁdk`E*?EGUþԫuE492ҩuH%^v*J7H-% }/ғ6scJs j.O\He-Y54} N 6CqF,?2lT  _Lpi#>~$,TctOs  8ƣl2 -9_床;$v .pQ ͻ`cwAֳm&pV(|N?|)n` t8:` M$ _π_$6x 󓆵ώUڟaL)vDDgԨĪqmZ0S 3`lF(v:dІW?∸_^jL#‚9cBh6:|sJݒLJ`߳ \+[UJ'jՌ\5 1=b2;SmcpPBas(\ "m0::sNP*EMȀz3B;Xm5Nd(Tk֌Wl9 (ա.fY.< O4x H=ۋҁ>&Y6#g; &A-Ēzp/=d4+8 d;Xgr4u/\G2xuQɉ[slp0d?q&O31 7Q؁QiuԥO H4;0m)9b ^/⋿U_Š95v>t9H[:ǬK d&,`]~u{星X ҹ׊) v̖{z/WDž,]&/_\5uL!RX%1E8HD3j8hRR9])(ҷpw#.t|p4I(SNyKp 6rJX×&1ncn\pG,`=LȪyMq=9כ#|tpz/j8)tVP+J_Z ~֩LJV8|BTS867Z!cOQƺP)%R;N_u1+rG#[1Ϸw#a*Ϣ+Qտ6Z{'3X{cYz5Md=`8!RVi4t{I(. %%a,+ޑWŒpN3/ v]gv뽊Fb]ΥاwJܥWY0v Ҭ_T3d0o_YRy  "o^|9Tܨ9`t9X2,JtBo9),o^l?Inl2 &&TEg-fwṵΣSFWo.buޠm{a oѴFy ;2- O~,NIl8/wQvVQ[K$>1ET30]p[@{q?XSxhPp%DE!MDL$o? f:NᴈtTAE5.rv܇Wk=S.g󂄮Ĕ8y hGfn5Ek4;\2t@,Ҧg]$8mX-t[́#Pb]!'x_~Uk[,zWh*so$?JgSj`Y }H֩鼿iyR9*x|Q ' &עAY#ivZ_K0'#؀}V( ? UچцWʊ% go9l,kh$?IkS/\وu\H&D5\0'g4Qgi\vL <ՅUuV[rDf G/]qwb/Y=gy?lR ut2A 3"xa#Ha*QghxI#ҡddT P͵Zid[1)ҩa_˭.'hj!;Ofsp3 }ұ}Cf^ uƽ蕪NiS zi86у1z8\tNå¤ kPј {/_S;^;׼[\dDQc~{.ꞿ`xҒ:{bm0? el)ӄ|2k68\BSpЫùt6LuKi^sՑKG)h"4W:$m 2 (6"cW}0" dӄ^*D+/~j&c~W] .[K[1UV N;lF$/,i:C@-wAԵ(k@[g e),nzXG-A?sqHr^wg Pt`kR I-ʴ:tMy}Lr|oᖄnɵ_ ,%1\0o1HWO{u_i^w鶴jΚ h;<(N-^~Idm~^ˇlrf__&+|R4,;cIӅs*S%$k|9w6WKǣ]"%y1e(v6±FFj,ߔ?wŬ!b$HnɑfacJ+Ь@I9h&Pnx_)B8C35O\`-|<|Ni0Ŋ)H"8a&\6 +Y~L T L]Q7v%wY#>~$,J@b/f(&%G" Voq4y2^.xoS,Ͻnx?~Cpl,cNgW$ p,F>Na8`NV0IπGN|A4m-^cJ{7LwB Bk0RE.a$IF!D0&fj  O)E tY W\& zyf ; s~u6mjגP^$|T[qfw+:pe>̂<(k+`@R–} 15RKkfw%iG/_b^<> caV|!B9," SG '-8L:VDž E%Ȃ=:QuD:۝ܛ LURyi*&VvgX0>z>\2nl+Fi $?vqX l48’0`_+xpN3/E-v{ds[U^ |HkӾ)"U)lUuy۲(Z_7޾a+Vc Z~ і ^Yo~s+ uЫ9a2ѤΤz[FP &Z}SA%E:~لD:DwMᶝ!'SEU+O7H9 (:p^n~b(FE܇j\FͰȆ{)I(&HN,[[YKƊij*xl @a4?mik0"gX9=LaUGTnXwo/kO:X5@NCFrF .>D#'1Jƿjm>qSܛ KvRWON0l#؀;+ ; Uچц]Ug0 ol+dkh&,ENG)zyKJ>dN^xhk ^rikE+"Ը\1ţiԍ⼍-RVL? ҚXu&^QN< Y\p`uP@Z׉dB ? l/Gv׾`랩w>y5½ Zul!ƃ>`z/G;$o!*:ʻ6 [ ^1ӭl.2*Bu<i9o8"NvW})k-p@mwuNtqpu(tmжLhu$2]qu{) t ]6'PNu>pbnu#3[B=y%N@HܖUW$'կw"A+C!Bwpo q }Y v xq&"%%%,U>aшaܤr*IR'eɧhW-qր5*xP7}1nͷU'd{b' C"Z=0%I`a"$gI*n]ct"ҎpJ5ZE3v5 p P1f9tB,֩q!ORa1B >ӆK(t1:jISv,WS-E0k:TNn:՗-OWtC_0[FӧpqU|6IbPJ_=m-b*b :0HK&T۠6 a33b.Vm.Hw dW)yX٫CG:$!] >) N)Kj]4<-uPAKl#h= * >yMn@y0/xݸe|)U[L#ק[,bQL+bC*xA 6`DpǘGq$+cU~4z椓δ颀ѵ" lHcG;f3Ȉ_F'3!lÃ)K|q 3߯uL]N |{[ET{dnpFgQ.T4?T:u3~iDF4p[ C GQ)gh8K魐L52QL'EFHrŧn{m&rsG16b(v2u?v%c zeû^/XXs%fy0!D$ߞpޭBQ4Nqzeߚ֨6 *C1BGq c;%0o8~9 u"|_9uܾg'q?UCR0Jd|T o܌㽌"g1ұ3޻`y{bBXo6Z"e&qkb]dwLW%"iFd Og k"Ql*:Zex? g |SW]1sTFӘ篁c0p !5&%4 AH.O`<+nVCE`/2WyTUAkA) ]uލM@4"  E ,fT첉ZBB]mF-TO%5N,xI^^+Tƒn,Jd\VžۡMA,2*ЦbBٌu*ەN LυS+DFGUe蹨i_DZ!](DF_ӳe NA#:ŭц,;$ mo 1mChGE3 7m泧K6cv(0z"+0"҂{j%gؚHF+= azV/BT-]O[ʜ)Km5LtDV7C$عHk\V IIh3dl щV¤Bb!*]mb83biAObl.DޡquTj-JOb,gܕWKd\[̜KMᛌsyڙjOV738lO&-&ɸ{yal:[2Mfͭ$mv$v ) iVBx˂ i}[*5\fwgzAPcnjyאcP罡Y|ճ WϵiFg5F%U~|r 5k_%Ibq 1Z%c#I2!JX$n<M}ʓ`CR!d&MӠd#3iy\ ]r>l6GNQ}@}?[CItr3)2֢f[+x3Lo"8)|1).IObHH,[+)E,3\K=H+6)q{6)˙wAO-q*u q}؍Jy0_H%MOsL+ѼȥSa~lv!q̹LlS3ڑ8hR8Ǖs|ŮD&ﶊ:ҠKH up#kF KUPc{EnbMݻQbk ZI0ߣpj*A]S'CN ӱ΃KCum/wvVnϵG'%\N0N?kwkf5TJG w@cdI2bsX'uHaz`G}(̊>YJ#*O֊-f?OMVȬ{eWjzʺ;{XKn24­ۑ~jLj!:հ5` M#}k5˧gm| $iD JP]Y̨hk7 ZAJfk3m~871CY O1lzy{Һ ߉ţ X`2’ÍTİ꬯d'&n5>ܱ.j9ޱhDվ͠dA+rL= CH٣.;TRQh57K7%Qׄ0xc/yBs$av9`$7&5_Ťߏ7x{&FSe囒hE Ĭءof-2B)@XBd-ŝxX]wg?;VЎM}kV铉HԀǠv ۇtoa(INR5ާawW. esIí'1E_x'6y0AABkS8<~/EL*K'`xQeDVVY 2*5i=CLmSP62Elqp9؀`(L),x"B%2r軓I&EE>6D` squܠEf>7;.d`AL,pR;Rpz(Bs[Ȏ`ٝVR弽vTsWI zDBIPG ,:1ُ΍(#="YdG$?8" j_" FQ^dTc#="y$i.gi{J oww7$So6c~7B(1|LWE^Q,YcA.MNvV%>e68$VE1IInQp9ܕqJs4׭븤# {+ucX dy bZϺ:,꼄"v:sZTbohQkB*ٷ<5{]V.cMS-*yD)E]VJvIsnx~W#!1mu< oB ]"4Nr|QaxbD\'ϹHrSRţy+21JfU0C$܆Ȭ4FWM5_6bn,o뼅7wշ[?l8Azsb0MxĜhLSp"M(NG}XH; 06_Ұav ~cy"Z<sV<}r$?,2}.6#/Ϝ<_vYoZِ"҈Kfb- e4qaT+LQ;HDbsjbz{77u[NZJ_%O jNA5qUޤbnD4y.htM=BsBZI=EVSeTZF("r*-\L8>}fm2CTj8,$!{8$ # ,+` {fe!Obnb}#dQd-RFVDdǛ[-oet˷薪232 D"tAȱrn6d5uǫ{yfb;!?U#m%ta[Ɉ 95b+R"[śE7NRj憐Q0n؃OD2NET !w>i7FHY/%rAXTR5/rkcFI&[*ML[t+ftp\CqP%$ y|Z(Z ~wZ V~+CkUVdA1l.S01`)wEhȤƷ %hn6]w|V55B k⮡>fğLH%HeB beF@(4k^CУ2{q#'WuHqh:`8$>ԠXI6Yd8f'p8&'w٭VۅVVD;hkx7EE@ fdvvN;|Bұ7z˻+Il?=].wݦAzPv uz5F5KL]UMD D3I_>K%3%Iن.sdCVdU%e+<*KQmS`ЬCZuHno #_F:NHpƕ~/C.^I'&"GT@Nut-r gگsGtն BL瞜??h쓱(6>݋i ?W2=l62瓣ǞÇ/Q}]'}xm4=;lB~{dDJ.laKx$h^OϿk]-}&!&*A(P_gL7.`ѰOEUsF Qj Q9QV]u 5VRSJ)5)EBD *dt4rbo $ǗGDV*S6aTJ~`aښ;)Xk22RPfC{|B!ZG qb)-dҔ6!4}h\ N3)Mp)]B1ΝTuM &DmB$lNKOß(dndäQˌ^DO*Ȍi#u@ʾ'dvRL2[[b: )aHCtRy8VfhRӈ D%˚XR GK[l.TRbGRF>2򑔑: J--)RRƤ)K6$dDp IxJ5n(5 J !gd$LFixt ,\ L" Ju%>wD7WQZOv \jqZűEA5N-J_w쥃@a1PΎb!?gfjx,Ȉ&@rjuȐ ql81d(Gc.n @Ѐr~$F f#~ۥ-+8F/ʧexTKpjv9kv=kZeQDĻ₅ʳ6f%^gq?rZl(օD7RHB"sDD* dDj)j6nQB`^^hɕR ㄡ$\Gj`lI uRe<^W ךz"E3ԁ(JK7Qɹl/ QkoDgIM0<ٹVDD-Ej~eUC0\ŃJ&`xf(mL|+jZ0T&ΫtU*f6Z'ġ2q|2'L4r-?Re"RmyVJz0T&>˞hAs\JـGzDK&ʤlRIj`T8l#.DUpMlgG5 :=bHL iA ŧbegX>lb$50ifn|0 4XyB&Zk{ ~I^- d'77-F{j`z*enD:7(UԝFvLЖa{%Z0t'Vޞ+2~eKGŒRRڞE{f-&Ѵy"wVRH@9>H 4% HA]Fޡ o~|3?Wq^NT놱Tm||Ǒ1\A2F-X Z_):uoILXIHP΢d:)7,`ܫEMe \`Nd8[_O//n0 YBՂB%y+9KKYg5To3XP3Rr8eCXFYp[q'aאyLq+\kc4,R5Ұi# yZgYR"/nm.ca' v &80>iNFj`Qe}#}B18-2YYi8Dl O|$O1 q/)ᯄn*+,T#} .QԮ7zޭ:P(h/ime%wH VYA&6FJ@y 3r zϥʟ '{T&7ֆjzu<"JC@9f&9Lfx&[s%bE aj~0Y }'GĶD(A'1km ('50 0-^It$ěj JsƨZ-ڳ޲vߩlF)d*}|H7(jtߑ 8MmP)Yn2>Z 'xViFQxSQ{2ՊSЏ}m/ՉphYUek悓2݋@m"[ƽ4>ny@=9km;ʠ;TYbhԥ,4>`\5EFER$չ E`-ʒ׵YIc$+[Dږh oс|VVF6lyePo }[<:O6FP-P'Ied],Rc=ߎQ*4ѲV[QU &𶇖F:ѓ;}?7 {2| .`,4?>sqJeRӕM=]V Ea ۬iE2>R/F $A8bx52-DBK(EݭJ>,Nn[XO 0ZCJIѐdԦ]XouRP*U4_]T'5-l(W,)H Y}GQ2 C'tDK-z* ]q\uRR$-:PdڢGE6Wڨ%HM8&x" :xCTA<@ YCl)& KԌTW5_UOA@(^,p.dէ҈%>B+k5lobŭ޳$ifR*qU%ōQG.Djy*UP=Nq5`}Jy֨:PLoƽK^<3Vw/`4HC2%121xir2nN[ONIL]gDIնirև6{iLmɖn T=ҙKQ,fT"{OI:9lN0 bM~9[ɖ5(mmӁ}W֋I{svza˪[g,]>Z508+P_Yv+,lN)Dž}kTB.2iCyh$jVR[0A3d,#]E ȺQ-[O&*T[^X1%TWڣ^稭 Z_޻Hz6 -,mR·豧L&ПL bs f󶀔UǐG?U Ǔܡ 1cRu sK@At9D 0G/U_/1dz0+ͧA?"^6`ߦRnk{R>k1-Z,0TiD@Gܢ 50ZJ;Jr v M`=d; (m_uPMiǷ-*|rTq#Rb&o&ej2%9 peuXu= q$SZz%0F+²q#&Pv24|7=][oc7+696ob$; nے^ߢ|;ܒA'X9GՅU5i>FVzgEifx!Zc []#iH7`2RR-{L oFnwkD{Ȍ|QPzy꒐Π5S_Zr:ε~ȊxWʧZ!OW^537hy+nXag:!︷y,[: 5?kQtݮ=BBlJ 4:O Ja41S:[A)T4nZal[ap@e!=p\d6k9x`l7 ̀pOIFEZ3pq}fp~C |'[D'[wX'Ȫ :Ɩ}ǁ:Mi"3T@zѝ }?w;amZ#w?OW*}2׿2әrIK7ZŐCvJjgu*bQ oZqR}"gqc@*A\:j%SrIeEK!, p.љOjݹU| [E=y`a(;l#]䐊Jw٧1#]­-GzT}[jG@ZF9èT{b4Yaڽjj򭆚{){FiaaQDJc0B-F R5DJ6jMj֊^qvsT,R 9W+kn@{NC5(Xc kn;VHm4UY:ӝI_3mT/pcuPwvO7AwN_hLם"`P+ylifE30fv8FnlM3oZK4/ڷ9q <e5覙5gGO ԋqVYkQE^ݧ?2V}BQτ1ݮ'Ywd5J,geul7F7~ieMXXD;$9i?mt"'{. τѨ"w%Ѣ'`v 7av/ڡkPNDx8v1 QF}|2:lW ssQ 8`CR_6]vMJpg>Tg>>;]W,~ ,biUյR)bGHj֢t7-~P?|bL{=r_߳ϸMLJtv짋.Vv?ߩs⧳:>=;.*V5~~\@.>v@MHӷpFxrrz&:5yOI*DY,w듑N|q'yw&RwAx8ӽo5KRnkNO*{&|d.KS:uuE Ѩuߛ^7u)N.a S;{؛U\jo+ 8%|`q[Ӹ||>M3kgWU I7U'%C`#E甇Lx薼c ^1DNJoVy#%/:IҀd3?VGzIS>2~Ċ1JvZa\F69 qR赂H4(td CRrB+tPa6aTUp>)OESg,/[D"/7"D*ISJ6@-P[!~-N'#ydkbEʖvA$G~'F%|lԁ̻d}r,,+CW@dYV QJLͪd?@V{@jD̅"_kgl FeGbKq )5mT. iiWH)ى%}E7Kak2g[(s) >{=Q珙OR~'Ƣ iDVī=X<1J-y3 ^ZݐȒ-K0,Eօs얼 K!{m 3/xޝӓej䕄m%8xo{[a]r CMiǛ) i¦VUֱT+}+JIiB7p² )qHJgG(eXngdŝP5W-^ۄJXrպMU\BjOt{NiD).nJS]AK_7M^K  a޲ɵ{nx&S{jҪZ1^)G\]xjYٟ1F(<Eh#]i'Q8( ؒ5EaPAبKv<P;+x`ܢiH6jרr evυt$wĜdTfgyK5fטݠxC1xYeCB٘]cv>V=s/򷝚:ayB#ܮ;,9\|dzڌ^Iez"FvAUЬUo, .p o`'xqW6v[ B4v&#V `ԝqL!SU>\-I]K(_^L#F n=fp}qLtk-1B\Z\{+M;ŨЋuǐF,imhjBŨ52#qlX~}PBG{n#z4h9_2KW Y┐`n!$lƺw7En>uB`#n w &7gjy`.p5K1.|485&wpɆ2 [A*xE` b Sl &+^#(FI3L 3&,jZ;Q& *$zHd@2Ԗ%&sQiWL㬊btz{y0Q3d8}(E C A53H JOsIiF_KH'd-Wd2FQ]Ʃo J2@Rp[]r !ÜŏVI se{ QpYXy䓜E~0;4"I=&Wi֡zV 홗6|>yͼɯ,w8Zwނt+זʽ.`߿U+Zxvt:2ޯwruJ>4uGկ3S#>~vS^n@~cg>y|x_ږ ^f:=vɭaeﮫ`˵{ jinVU@&Oyʯ8\-ɺgv,~qI1ϽSsGV:b2ueVsKlS1JR)> Wy 8g!,WdX[DmCԤFI%gzuF%Љ,ѠI8A zx0v#әX G!JBxX%tzDgloнD47Ç!P&<$jAgMZ"p{| E5x1=)v]k>+5b{Y<۹hvۛU\r1.3>|{w%vqkll߫7tojCwWg΂lz}h> ć\_J+cW$7VTJ5 Bn+s|WRFv`wgkB޿'Zg;#s YIszŝE皎Au?LEcwW=8ALxd%\)>7ψ kAsoyݨWx.y5!6=b beg&{*Æh'#^Zj}HeԺX9Z&oUʆuk-;fJ_G;5Axun%XPZ *9LP^%]+_UF̝saS;k{ENP\rč\GK$*K0ror<?u6j_BO7g'%22֊˱b͗GHe2w\mxŷ'H^jGU )0Rs*܊ej4H_ޠblj$໩MQrS.t,F\ *xZ;XuakxtsxBߐ2j뛶^X6[7'/acHm:qU+WThW^r}6$B@7"` 4M[sBS1"pȹ9wH_A]uOŷE,CpXzrh-R\#spq$DD9Dk#cCڠȣZqhg{uQb vCP@\C-=:(S_ q&bweu[qRt+j!2R&M-2:b=R1:`-LkC-.hh`H~BSљş.<9~:&0W.L9gD+ |/')*QT?+xO|?gܰ{smQu9){nyջўGP=xğ)3 }{hčm|s=Y"0I_ fV1p\kkH+Ny  AD ,5GQ!:P[}.4g`t{>VO{9zaREdjNZ<ͺ"'iF|#F/F|JF5*b/KcRAK Rn4RZ򎲚ߜJv)=)%@jֵCBS;Y}--[Re烙 yev!u&o68jϝCXV~XgaS^yQ#TLUN"PqL[-` z;;xz_{?}>}8TCBhtsŗKkKD-߰qNJwߪpnS&xm=8q?9z2*"xb?P1 cr!0 {\Uhjϊ--@(jiXcElp$*qv:HrWIn+쳬:xȅ ܡƪ%J6Xjk] RZV{8BS1- eCpKLFe)=ƈ0fT{{|T84=*ǸFcj6S/5v]&)GO=5014q @Ԯޭ/bBtۉ?Z?!~pgdSB.r`H`Z ^gۆ @ j ,KW[h% 1)w?n6&(S(]QLM1c-![jv.<&1*䄞\PB#譸H'Lyl4&l#gSѩrg6%fgb"*d.>;Ul:40 LF i6ZQ"4F7zVIN[m%tA+-Mv*Mm!|WVjFU3:eխfȂ,6ؔ'[P A%(Cы k$f5n[yјi^,gN8D=#/` LBjCpy+ա&CJ(XшrT~z(, J=y sGM︬k/\6_S}FO:kj3qH=ecbSây //m'jsR"+6hM$vşի!~CcZ9g\B5}w-VNqiN7Lrh;VXWE*18h dL|%X{4Wî]E']=a=UOٌƐr%^VV1(T.J9Iʉ4"Q(9u<\EĊ6wI|&!`NQ=ARUo~PC6]KT@\B V+ۆq;$`n5njaŌf9arduR^`(4zEJvP1F,l+R& ˯+aֱ uM cQ7x4EW2?Sa$aë%Ԏª6 W](wI¢_0VxF:TbJr/goE, vX֖`HF;vhҶgoӲ斊CkeӤ̚؊ѪIԞlClJ ١wrϘ8MVAqRTʺY@P6EK,1]ߏO>|<:>5+Qb|t\&_H|5#jmh;raJ񦙷){ ̌af43AY6ۊkU*uՃR?[$޺TE'rTsʮJЄvgn^-egMQ֏^*?:OH *rn, *TTTw/p>PSʝ,H͢(F=EW7n΋,NU}Kv(\QZ^@ݛ][,U?~}o"xl?(ɛ݋ֹV.*#x9ީo7ܭE[yZw-{-}wQ-ōY]_.F=)髱k]sy+ݺ{fp{tGCn0^t8$U;%7S|  LO7߿S^vیf6Y&24z"2XDՁT*y$ML yEFhW)l@MXCl.jw/0⥤k94!z%,STkF mg]䍭jmS7SwԱlW5avIH!tT\GbO"ʿGʚz9Qy& iNlf3>hf`fEVaEmf憒)v7[k {O3tW\Նux`[.TIVbxbcOٍpQ,h{-s`!)հS>fg@Ыk^pX«hg{BTaB@Wy>Mh6:Zl]\9w+IP2,mʨ~2L޵,"$u_oX;8O^Qz7%ɡ<Ԑ.c&3UUW]UIX!ֽF,4KI0g+)KTcܮ6QxNlVXR N# BcȺR >3pF" MҨ(v.ޗ[3iZY&8Pj;6V[Hx5(V}2Zy$M5˲Ur9EA(IZx⠃U0R;L5 Blg$`8Di]ɃҪpUa`DL*sN)>T^Zsۿ}OU7$յ8V=ۗJr)Ar~Es&SN) jRiu ?BMF 栫hOEDLŨ"HjKBEq4,$Dk*',5պf\"&;uI=J֘ڄ?]Ŭ6р%Ҭ.xP ec:R԰.Zo'6H64ъl ]HFWE2H;q,PG8[_5 _<9]OhsF*M\\FF#j89[&I(Nw+gG!nibճsԉ]\82tGѝ%ĭa[s-y>3LikHmk/ o爋gص${f9dJIPTvZY]mq8M9Suys'+zI]#ƚi=Ib6֛> {Ȓ؏o]9f?Η?ׄ<93&N_F5~e1_3 '9~۵qZK\ihnBSD}gc ?5H@ Omza]6vC¬8?=oϧzш!H05[\v?:-'}wj #!bP3 R!LO6F|~X}Qk?~ }x]A LoYr|_Z} 9%Kƻ}ZֱXCKv ZUJOd "Y"Zr|"A99b~ "^5TV<3=dw}>?q_(FK5x㵞8w㰫|WU>RT}ԞC097B\=iĨE?f0%Oz\Vop HyJ8[h}Nf1_: 1l&{a]lzXr]?FgZ\ه_;1c w knި%IRzj\cڕE{9wV+ËOfO~-=$VSkط0nTpᆟŒ͠fF_*z?_ԃ͛,fJFykY‘upʵ\[e5ge-;)0Y!y:REMG>Z17'PKڜTZr)65]{­gNF@hυ;g^ c[5]=K}A~WH8ἐ=1HPR)9 7͖;L8\ٛY'+5ҁs*C,HYک*{[<ㆽu,ِ|m?NhV Ir)gVSFSE Z_-/hnzŔ< 2}Fu' |?n@ߗ~n'νSν:DnA~Poļ(k&d[ [6x泺#!ϹOoj3u9xRHHb=QucK}nl3}xC\'FO 1H9%d?Z[Qi3d.- ˛^qf IO~YNlj\$?kA;;hF:bLH?{\\ cd}Ah}Uᚏur#5omNl-77֍'FBjETI8W?kO䥚%^^F()NqwWH?gl&0῍t"1NNC$qyU*SY/̫߫ 2zhh}nͻ}HPSOlJ*&U^[i#)c{D\_ ԁG⚦PyT7LOnݲ{ÈJLhh:]߻˅z/*Z?f];ޓ`F:!ݷnL{2ۓY;u UΥ1'uoJ9`;*{&*NFu*EKw o֘H.TŽo{گ]ƢQ{333}CDzX  lB1~kʎnʙ@>S:UܗRdCP "Y+sEQQ>9% ?D~{'[!w:S;?Ḵ$gP {p$䴉p%j%l0hKCH&d,T!K d ov0^]6?_]1 h.gEUY$[  xIAm$"R>O)iО9yar%2֣Dy00<V޺AMg 0_EwZrn';& ,\{ sx&Ҟ[ѵA"YG=1 ع$)NT8hMH# :#sD-\zxx3ƭhVԯs~m珜?~8o-N!8+矡6W]/x\+J37W^_/E>?uo?,3\ADoLbGGBo`ߴſ> ByuS{wR'WmDt;`${;NjKE~>`{īgkm_F}tXK#,WgXwm5t~1^;5;?:47doCa*$%.wV5!FɪdF)犳Ih㤉E/%T >8%o5ɯdO7Q6Xk>V5x /(RXE\#~R ]n]ȚϙՂy!ÚGb%8&P88-+>$lU9je#l@beG^DS/@,}$m/|yuVJ&;$A2cZȚ k@P.ymK ٻ"4d#<>q Mڍ,$`)z{{w0k SmK dmS~+ɞ{Z>eFS6?og<0j3W,ԛC!+!uNUN޴M|5¥T N(srIU:ƒP` (Yr[ ZѾ9q%XGCU'7||y?Xka 3/ާ/KFl@WïcE<‚ ;*ww瀴%1bJ4*dR)DEU˪}UTt*)tZ\T@hRI!Z\OCAHr qëRѳ+ ,(lgI/_*f'917YҵT/u[ ؕ7o>ŷ+؝W՛YSmϻj^r#6{۔4ܨn=>EwpO'AKƓ:gjn{SVbخ{Sc\GHsfN 7a`l  ½1:b}:~Cd(ݚJ%rdYͲu}ϳ΢~ "q0m+[7xUrz92Ÿ樊 =+OvY\`qnræξ?VbM BS)+D(=ڣU\]6MIfD]<4%3lRA!4x{pGmgBi@dE%ןiczB 4=ϕ̘z)֟w j WWX5I4ގJ-(75UsojM]kA[S#DJ/VV$ 8,:wdRF*%XH&,5EO5ۂa腖a"]u̱5?٤l/GDDp~=8vmM\_ K֘95GԛH,\V ɻ7!HgdE}*PrΓ#eKl86j:>>K,Iݤ8v!93^ξ]]}h'fEM4Dƹ(z:H\T9]4Y;) jcym?_"ĖY9lx0~ËweŎZ fA]pA#9`Y!;pnF0 l<=`@x0t+no)zH6Nϑ[ Ux}X9BGCE/-TD7"!/.Hcxm6?k} VfϞ<CC=H zf.F/ׇ~eW+]_~7GVyOދga]>t/cE]y3ƚ ]:xv]"/U:MAZu]!^@ciAI.Esnঅ6btwXaۣBzϏ?-eg KÅ<@ O'ѻ?:0#&5\k8t3.DG;0A[P춧n(~~N>FY.-5΃9KXaܿz39ECͽ7fo@sx7P精 `)|.r /QX2.Lqo c6<8Gg$?ZndbqV:>̽eo(X\tRbP\JkڌmآKj̣wk;Gf@a~E4ޫŦ ,eu^2#]0/)EbHK{mrie.ο|Zƿ} k3^Pk+UFK݋["׷ k ͑ȧF^[H~ h+&̍H"X ;\_Ss^"\ Q.ϰ~fhԡ1$h':/g!@|vw>M+Xlz󷇸9]xL_/Vokb *q "cj=Y=89[#g`rэq5o'""}U30/`[޿uX?RGbC77z!~#r[Ь+-y榾؎gBhT(E+$cD B sz ?JiRJL]Q}muyBcQ-&H|0u\+ZQHKh cA PI|vmU7Px"!ΣB"h5dcEH,rJ"3EUh}V=J T[KKU=!(F9\%v&{05 I/)47|Q{s*v)|s\`%Y7-Rr(!fDZ5.+9&H7ݲ$H^c}u Bh2B+恳׾'pTM폫_N.~B'fs93݊ "ݺ t,E Kx6zp{D$6zKMSCdjOk(6{2Y=ba(qLD. N & vSKu P])r/}BW#K]_I<Cr\_]5; |h+צ,-0AaQq~퓛ֆiV{?.v o(0v ގ` ,b$ Í ~scW=O}ZH@;$*_)|a(:V2r*۴x<9AȚv n96)IȶT`Ȗk+C5]5H*Uǜ})6iYtfJNU^\?TNu;`05q7C乗D>x &, xn?CDU=jj5ը:(ﵗ]Z?*rh)KDNle?AnMT]D6rFTLc"櫓=6>~^ҥ]U6!j߼^prgz7A ^Ew:Ëي-6f ׏ [M a/WLkmH/ ~w|\KfE~zh,$kUMIM3zPdU]U,٦B)TTw?l&941a1I MT$UAEB)njwDJ 9Q#PIJ@IjL J%?Y^,apŐS) @=!Dş TOVn{?#7~0.lRCc!iॆ :cWyΫ-^ES9J081e&ֶj ꖎFIz58c i 4R1D _`԰|9rے=؁2D< J $6?(h)LN^2"?|!SR/OI(_m.B7( iLؾ䕧Id㺇=#IzȾg?W'~q~Y+\ qY|yY>6U[̦YZO=cj:wNR=N䭽B1?F_*ni)>?)J>*]&!z`j{{\g8_ya&/G#a7ASG<(N˭#sjjO@9'J+덶0&QxIaR?bd#Ѓ3:> /;69b"9  iDPNlSI`i|S @ sM3c˜2\*-10`O+!g`>XM\.O;2ZgU'VVE}_Qɔ%)S$QiEooK2%^ sI dVPq:i\Gf1s[ hfDplLټ* (MJ wjAeb$@iDz|fɖ-fMBʚÇh^p|Cc3D%{"5%J(a2(8[C76ӍSƩ{uC7@O{W/8c)7ne8bO3<7ܣ')ܣu1'T(TC sF57-Lgn!K:Ke6e&,!їۇ4yMrdϫ1T60{g _U*'A|uvW)-[{:Wrvqu3~u9֣²\W1!Bf}_WW;>Mcb6zgV1trU.GloJR2[gy}1c2&P0;Q[/g^]iv(^vT/9D^) Ȁ;8@4*XC?<TIֈT$38mBy$j>!!%>X$L2Z6<'Fk:4ZeO*R&FoWvk 5$ @z𑡼֗`|Q6ˣchhmy\0y! M& {u;I%(%< O%XSåwĚB0CVS@^:7j( v hsEu";NEg7,Vd~GL۬npQOFvd*Y,݇ Mr|o6'cZNƌ 3M)Ms[$~?\-z $g؋{;蛫J|E h+#aVӯ=F 7:%ozIDrVXj'?`RȃӌK#x͝ ׬){C4RVYy}-FQaǭu0[o20k)QoBYa,x˵c9"&XmUL M B=1if̆:h(eRW}:F-"m5K\rF=!t{a"J 1PE ia2JKfJܠS{k$933'(ݾeFtr.eHwIZm{-cy @$tQ o w7b_n CYĥ]s=[,8F~ o'2.ЕY|B,{=b?R=tr8_[oUͲHdzrd7Wb/n_;Of5h|HlOe4~An6 ~z Q],Z4|K,WwL&ΫO_}X܏~mx3ʏxn\٪2liYq]M~RW)9EPm'Ujў޻˭SJMa ڶGQlG}F67}XF}>y˭絛cń&3o+[ˠA֜tsT+kR.TS%rԷB]䤲*i? ԿQގV 7/{ (_E?G'WH-B.8)1dwbtㆃ읾9C#FZ+/ʻI;]uT+1t s 賲޲ {TX{z@he?KA2\ ?|zKMb3nOVGxE˷K{v\u4/A]]?e|ut]&iou1r ޸Ӆ"3B6BkOooi/@褤E"QّD<&3SVlFst7:d'!E#Q>kS[_yd5{%ׯZmnW)wUŻ6?̷oQx\|lIONa0-X w ꈂ("MƲJLd?{W6vvE$L}ulߟٱJ'Շ"KoߐÙԁa(e,$@Y$1_<}P X\̩3,rR}ԗǁ֪—~.nfۻ:Ò< 2#;#@Z'0&\IX͓,VKrҎ}Hfb3;x"dNn2WY=ٞKf-9ټ E7S&!) q3k*$eR2ϑL00*k?&őh$6fvbh>}_=QichV@phA+# 'T;aT☲80Y!c%0B p'1d^;%îʔMI֔TFG`D 22uasR[)Ȟb_gޫ~c/FW$;?f֫8\%5m)|]ZB=+}B\>i Šע\(g[K#ƅ 6NwQ^;:Utf (6y>asM94a s>R?K[IjP1 JQO޽|P`C LC;ݡK!i*GZ-"G#^g(~Q} x4ir,Yyc|fq|>vu2cUVN}c#gy{pY̿f&Q|6esqEpbRE||NC`%F$IR%x:}Sf^d22kk/vfC r+Vm`>~e>K~Z5UMPY! *[0 SJo[0+Lh v聉m `AGkC7W$ F2kϝKBЃf롼XJ5lbh=FkRzp2߰@XXb*!.Zh5;lGF-T qb i(YrMpbq+sG]<"\F NOr>qT;_N5u&l_0\ZA@| Aٍ5( )[bן)P!UW! Ӓ( *#@* 0.1dF}sOZ Cd G8Ylg ļLkĥ@U֬'U. fidd6 Lab 0v,bΊ(0)nK ,ƃ1/BDB'B^5tdt9= XDQ"Zh$|d vҳ &aUCP)=ݹ/`Sڡ0쪌!Uo Zc0ݎvI"^FN)&_b" Iw/4xVikiIJzю=5x .jzz^K"TOd k2:e GZ-unh)Y (\c\{v]Lu!v][GƝoՊv/qyOIw3"dAvlhVTYDR'ML1f"XE#K%IayT4eቚFT+bWjl0J>0aN`$Ņ&sHV8tR->JȌ!A;Ufb;#bT` RB\rPȝ)ER>PC*!&{Ws-y0J.%`rl

RvxM:+7CcG.wXt9n]Aڮfш¦4ABǒQdd,LYͬTsil miGu(ܑ{vmZ6V?nd-K:~lZi:TǦ&%QفG`E.ur]kݗa>)"fViRb{+V)X]V)X*E˦> zDॵvT?so@&c+TQ i)yel"ZI%x-A@5q`Ǒ (#M etc)QL "N&;[>>o>68 H=!)JwgViR IyVsd NbbI$pV-9qEja?ob/A9b >SW*{`s%.fFsm "h+Qު0pE9mly\|3&~t˻"7d!B͛WEpk?mS~ /< ;x`V|ů<]^ooq5n,`N@ku~hCQdfLGN=,Hx4v>%3~l-n ^l&1e}w jHE637Ub,hqr~-Ӻ>EL\ Pɇi \- ` O4+x<$l4Fm9\¦kk2֟XqblsͲA40hV}]E8؆,s]N,m#_#4<"oḋ!,Y<}{"Y<96Pt)~jtޭ'M7"NrH/(InmRj;ho..\Ʒƅ` 0"4f}5$tU69Ԛ\upOӕo++ f 0 (@8Y,WX. uc&$`/ypiUoN.TV4}E$ @^ IXVU.{}"xӰBp)lCdGZ{y<#M%+rXCԿی;;F!d[!\>Cnrնz=Mל+IaEj ^r~Yl!Mb)-xZ ׶d=hd 7&6d;%(BNFKM0H浣U6EcR޺ϲ|hR CsH0~E(\wW\ \CaTL-!kS֨vxzZkVGDIK13hlq na2WC7Lbi6} o\s$D4 CŖ ,]ճcP>Mac+9gu n f&~9S3?; 2'aӲZZ3`?NDSdLblK4Pʓ'Af- 2 v4 <% ,RN;Һ'f僃*q OV{2u4'Q&(i d^$'ZEWAɟu:dUpOCN@S·d5786g-[.yp \K0겊᜵ih)j rE>: 9?Eqnty% ﲳA_;@x&0,c(ې{ !_Ͻ-=VT=({_/g4x5Y>yg烼%g3YÓq׈w=|')}䄽w4OAfJ`Kd~N O< ]G͌_[#ʪ,ՅeYӬKO[#U~.6r~6צB w!U;uA n~K?/>buhzJmD&;w''aG * STm{/Żӯ|j)iǷKX _ Ԥj@Ƙ޺j%YE:!ddus"<| qm fiq6mm Bd}P|aBиK2,W RmxeEr1j57޻xuwNobܙ<~>@=N-)2pȟE&C*٩8u˺lN+{o/f@4p(dAcp(uȡĔp@d'AS("_P٢CQ$,D8СtWӷgokemp*FY*^/׉ZC6XW2ڄNV{v'˦j-@Zv4CUlm>I.Q`1"=EOhq"ZO*Z;I*~7ᷜRv$lQ!:B]՜e Ze?J!3Nb2 4ކ̉0,gu)%'K:A &аuJ={=M҄I>R3U^ov+`|DF5_ Wc~<[>yۖmW/&~g`vP<{o:|C7iVss{ Grý^~~f|pB\f}&n!!CB>GHG46=lyxs׳ `*/{sq(:;37Oj릟*~6g[3YH3Ca޵m$e0_n~? pq0G%{ƋWMR)fS@{HUWuuXp%P:?նqdVԧbl>\E١aq+>1h\-<7\.%4 )Q_is2O)D%&i(()P{a8FT _OMPe(E~-Tؖ-"c|8ٿ5f>Wn!볪 ufmQ4|vy7{#X%afd2z<۩^+onMm cJO;|zj6 =3.:x语zSDHi0dcҌZI@s"`*Dr3$&@33'|t]6g}LV}Dm^"v+5/MM&m#D|"[KuM}ei4'QYJ0}~X:C7{&HX2 VTR {J"{"/B\}^|$Zo"V4MsĶMKLӫ:ҕ(|?.{m^sۛ/V37ooagSnjxMfm:Y=(1KdNub6V7鶷kcۆ v2Lև)Qi&-M@B& <`O6}@xIΧƛ7%) I'Y<%?7nڶ'^#2Ġ 8b *[ӿ$$z/˅X,sw$aL-/;-IS)='I~Uλ(ݦ?4G&hi##@z"&\\@؜1'w(fQS*/j>[ۛh-.s/J[alv5ʷ}nmg@:`,)T`nh~?x%UaT0cceǖe(D) Nb1T15uȰH -cpyDHHD!<?1 ,CLDèvQ&$^M.J6t)ce.hLē*"NW}pANc 49sOWL7uFW(~ _Wq.Тb |aHO~?@RlΘ9u dDUXymOc2s% "ZdQfj[jnd15;=r柞CVbJɄhv᧝hžd $*Hh\0hZ9ߒh\jF SgXk&gsИŤt["NV ;]ziʛaL%Ra`՝nq6s L N54|54w"ce%'>W N” ka2G{14i"Sw*4A8}!.-%wUc>^8'QX~oCݘ!1f=} A}Ӣg*RǬ7;N tvGf|\<3iBXO~0-?}n]C4>Ɫ)}=y VB$};SoYl]*"uۯno*-N3Q7)XpIsvB *fvW"X3$׷7ǩH>dq ,J܂fKV&iȪzs՘WUX#+^=OӰV)ʅ`^x)[ xdLeOW߷* 0Jd?9]e5 ׳hff[Jj&~TdRbʛv` cL6/7ƧuY!2i qu2QDeţNw3,q*3`r޽ڍ3)]M# ³hb&YCU JJݘgW7u 1RN#oz#^)&4jSweD̛T5s'fl9:Y^*SuGJ!tE _:-gyNl`,7CO6G@|X{(=-hhN=Њ H~8fY\\:# ^LdPB0cO֌8#Oq2 4=uDdg&38̸ bJX9A9Ğ)мGŸW ; 4oi~-SJ ~? (8 ي-%H8҆J 7& DthQkGP ى*jG2<(6U?o+X40O#AKM$T"47BCOŤgױڂpEP^WX_LֆF&N@B Õ}o-H!vc@HJs >ʸ~ LRq7㕍s&˻]uXT/~Zs_AC<1F,H B B|B4O>6yqe?|Q`(#[`h t -Kڡx#R> M Gī|իeh)!>[.(Dg;?rct) {fwZ~RqQ~V5uk;_-9(wQ3WRp;WZj}a!bZ(J]lݏ4=ͦ9.W fLyǫa{iwZVr j0jY*(qM0!٪NV !׎4v78^ @Rsh;v3EI_1U>JM`U&^W3G{WO],`W)++աq?FPz/Clľ^zfsZΗ+t޵;Svʔk{%=w]cp]KS <?uV=NS;|Y(m!EV?Dߖ7&(l`17ˮF~gf(Wނ-®Ogjj4rϴkr! ,? C'\pk]S1p`z bWkL,k]o\#Z<;D(hg1vGV<¡ּZJJK2I#D AodQc. жh^龘|N1y Wcy-! cGP g6 9h,2Lכ/ yOȈ ,KAtt_]-]ZB"GGM;LJ0'{̼-3ҵ= |tͫ1+F@`Olxl3.1DN:5nITa{?m[UT#*{qF=+}&6?$(Ї;zFQHW^vo. Ev-J @N6,Y0A+KC[11cnz#'㵋3N*oSpwLc?=ZhxC"*ķ0d.ȴ\cZ0@V&ru91&$E. ?F)z+bt,ڝes; O'qlU-ch0r(VuC.\ =?__6T("3dO3_1h!8gӵܝ4'iHS >EZ/DKNOD0*u?^ƚ'~9N->eܵ9Ur3be;8E+ʌsfZ1 DrYS  EcB((l[G79|r63O|1u V"# ܀l/W\k[>#sTvQZhAKO,B-+,\8( XXb2@q`2gڰTbppV6gc"43m 'blp-|YRDaa~ Bq(Q'SJ(.zi~%U]N7ǽ[צ!n ^_ % r{2GY֐6F 3 phu,g Z/ҝbE{o8|4)}.6`bsJj]lu,lXU5R$:=[dY ] :$fT q n%)? o6XOGt2;rUtGMNv4@asBzwT2ю|c=9V@@s[JW\ %=X@`KqW\W/!4AzߐS%.0H)7th:tT_ 8k=,7gsY) f vB03{#$ɨV{ %{0!ě-QFA.+[ѩHx,Bj">7QpIiw*N0X=Dk/g;:ڡ2pSHW[-+`i"O@Q:a457pX(b33N0 1XdiXMGಳ (:)@=D:2Sٷ\{Q/ VpPnvl6L4U.w)eZ KuWV1H Uga):Uk0{6nm`ײ/zl0n N.Y8YRQ"3#ڬ٥&*֚g9 O˿PVlKr 0qm|T3!tpHsl@Fބ s^}6.U+AZI_L*SB,9oXނ!|Cň4H#=R\%'m<'Y2Z嵍LxV@".YTR{.?CD pCdz;>fAB31Ba!G@hhZX}~5y@Gu&j{!%ㆎ^~E,6MdX$ՖKΏEZ{vRgǧPSIiBi :JN/bNmպ*Rt 5?yРc8.2՛W3Nǁ:wMtIAKgU;A/Y CKRsgC?~_+5gڋחB*J} PyP]bÚ̟f7^g=;LCLH(!D2| U:􈄊zM~{N+P_ν]VUV-) l̝&"HFp֬p]Z忓<skx}ІH9.jp@cMԭcm\%mh.sj$ QN<P vv^G&? &g:!/@hz}`SK@Y2Q>Weoˆ 0 \p"Xۛ]0hL0Mfu>o!\/f`q!%c_}%a"O\4Y&Fb}bMJRx{PCAKhο6UMDx&ELr.gҀbQ5k6mjn|#K^ R?*06.Zԅo0$4a6? ܔc*!@%!2gx0KN^u "+u 8T, W09 ==@vLC>3 Cd mO,(D l3u A#7*nM%nZdy(F'Dr9UDX[BPh;!]P缐S9izG}`3Ιs$s$>|ʑ0;)˾YZTH3_CW$ @&MX~*sЦ19xK zs_N}B H 0?|J̒llgU:j<>0puSA04~8gڿu_JL)'ӯ$쓘Jw&b8ACS| bށ_$GD[gX1H"> Aؤ؈yB#$AGx@BċnㅡE2li6 甘0P^:i9y'6 )k{tAtPj/×e2?cy&aH5=ݑER׭J5y}gZ |$g!POi@ 0\3u#]uU\^wt}?U&f#n޾{ۗW۸|歯_ś_^zww}y^*MYuzIu]ݻ7.g˜x&U: ΫD;˃ן3<&* zL,Oľzk~ӺEqvk:m~9 ®|ȏ%'vԓn8adΉc7i AX\K#kr^͟~?ӳL?uOf3}យ{2R͐< X >5c:>fWi~ЍqlxI\7}odm,Yduz0_|lA0;׆wEu3>|?BvTF6ta.z2Lg_*Q p*k]͍uvZ>Z>ZeWO TmF;9v"=?C Y{ 1ݫtN}^8q&zzw15nל[7qn6_hs{i,<;1Y~@м7XCFOV=75If|gͩycʼnk &b~Q  %dt)3Z;ʗW_NA\/v2bQ)X ~ ^W'+K9 iGH"PnCd^@IR)T2MH8 *ROQ#֫%\XICymF9d?J#ҩMj@p.c2an⧸Q; yCضqHZ.y >C@g1E§  TB#!T$M &SMQ )l4 AsI5;i7:ˣ;6y|nmkSЌ=˄h&3EmM,lcF<}X$׼"ɵNl5 !K31ڥwQ74/{Ȃ+<$E {mQ ? *s-clTHCc!)\YJl>aEdAb*k/-ZdKR6C?4xRQnQڕƹUj}Spaƻ|wo\`/,T KL '84޶Uc:>V,Ĥ'ޘ$3la+>Fg)53擏Xfoe֭Ϲ`<9UjM[>/7QGqNS%Z8PH7W%ۖ\N $~vsA/"v}0Mby_Gl{Tif&hnTxI~W$bcLSQJja"GZ+8Tr8H pdlJ?HhQ 0ٿA st#E8/s6Ь"@R/;VE56t}ɯJgyL^]NjY;0Yo٠bT_u'͓`c x:HNj[]5 xl%alvυK`Lh*;DGiYRS, ¥uAFEF@)J|T^bLƋIr@l2ƬEkՠ3bRb٨P Dd2ƌ?M'Ď@.~nSpfM<,%Tlubl,VnVIZ4Esqm+.Wl ~&Ma;-&ۺ7A(8EApWGmy]Y* MێѓH b1T SǸHS<5ijcv&ޚ51Q14s*< CXFdR%;NoTa.kh嫛gqErf=wnBaa]l!1N^ g]λy=yOP7{MM-.fq|#9HN.!-Um-kk9kP)!]u\66>7gHK3ĴԺ *ycD[ 1;$P6*2U"we͍H023;[CzPzn{^ H)x `fW-2@~ȤI iVa&46_`D+G5i)pHf1YQ&*QZ_ް9l vZw"@!z7˧kqD%#^+ 0X+ L|J.XP*L MJf TAK/6Sk8RH^rj \T2jbT׾*]tY8C2%dm@Rga"Tp?4e_L:);`OW|b{׽yQӛ Y5}RzY]Dx-YxJy1ᩧ8s6LrBA Gcoǻw%%Il/u@!H4] '>LY7`&~ܻUcf9Y k,?^`C, c4E'ޜ8(Tw IE8^Rg?$Ha{2Pg"][40J:X7 WYtA Vˊ%ȶ7x^BϙqdB5)WE)X 6J|DHw(|Jt(EoavKR0HPS=2#N#BRB, Z ;V!4TiinFg|3˕ &Lhޑ6IP İT>W[N/YoH!sXwbD $lNr؆PY͂Ǩ6u(E޺Q][*{7&f;z$pvA˧ n=mXs6[}dbo]\r0B6.@<`= ܧp^ۿv̋_a* [>w{׽v41iw,vX:wj'1!Ns# 8J$()PsL4( Kt,W"TqC6)@ĕAM !CI1 H3 )T^1" -Y96tuF%V٬fbv׭h1Q1i܏>]\$(0M?w|. b!vˊ z,yiC&;I.nY{fOgc<"X`[w`t~I|5/Cu0wͿ'b4ϻꛅ K޵1=*ݼ[vwոNf l}5|Q D!=E:y1yRY~O csOy. -NaJj)u4oIMM;':Rw5$_/k]!F~a@+YxB"HL?2@Gejb[@V $!ֵy*Pstfb|yNP;Jpza6]`cԀCWMj?60#k2d7$v88^VGleb'bo6h6S j<4G?췯www[z:G OY`"M$BV/@sPʘ@ ) WH&)C$)f !_42|nf:%vWThCWC|1u1zf_Y53H 7ce/ TVo|fHZU()A HtL&aZhoERJrۊU;O*TmGk 1'UEk+ z"WREMZJ%Y :D/t8B0 \H"ߥK. ʜp ܥO#G(pN\,sL-"!0<+dj;%ǮLmthl+TLw ؿE s6y$]Ԉ|IASTRTmd25R:mAvT.沾a_9Xzաh\R膠RQFJky]rtF)!CTr7W *[x%ܵ@`Opnpkgy!@%ʎ*_VRpQ{]ve#y.9v6tF.+]3uOjQ8P'8i:[̫TU\n:Xh6i%E~4u3ٽ~/2 S0n#1[[#1ql*Y]_GJtzםNFVv_u{<-jϙi/u>Vu—;Ѵ-Ph~Peϝ@̲Ya6o!P،#{O[e~c;|m>Y|BP|\ZOw>y%נ$ MtǦMCvHg|7OIB@.,91`=[yo۽d+`_U$D沓acyCnYsvl8z>]etVS'L60e7P,j}&^ٌl=gR4#Sh9_DP_(@B{M nT3佨"ˎ0 .XtCa[ga3bp^xBBNcHkİ9ͫpmq@V|]R4kyakNvdkPQt[z(,oc4DHx[O"PT|u+#IE&+ %Wz󆃕PSxڃ\&F#j" M"|)w]%D2HhAq,c  bV,5PHY#9 tLybڡ)A( $ƱFaeqi )5Pt9lN^Ųa9Jώ'9ˡY$;F,=3+%МYK`v[1سSĂB3kmK5Hl30s ͙3+TL9˙$eDUtlghjcE",$=G4לʪLҜ/ (wx^xPqJpϧa1)14Rz&beV5 !$ʿ\ NS|8xRBpByp8 FahbhyN9-d)JS?5pZ+-vΤ ǿQɣl׉A"EGlG NJDR!D+`oA4xLx6w12>7v0K,m.=HZYJi#`PrP|u*d;ʅ\Ua`|PhU8RF{Q *PKS{q01f!Ԟ\G2DoGNIPB˦`VT0DVQ럅B2r[K~QP첂)0\ CD. ߅=ޅ"SNν5.z4.i;FtM YSgg&)Ă57p$ϯv~U-.2'ψ C~&+*r&Rn&*QP[^Ot|F\8<;)gkjǨa [^UFz ah ›VHq*|HR0vp=X;p{=YQ?];q ѵVJGU>(ݬEqjfK=*pe}ۥ/ +wyj۶Jŗ`U0/nul*U)*D-q(;vuhLW Q,jТ8[܁B* : #"}^v{7ȇ [ΟE*F2 u@@t $ ĆR@*ZP9]P¶p~m#4":nz-rۇlr!<1L+NgU9u2+^zQ]/JB:g ;aXe:e|&;eeci=h9(4,+76%e.6==oe$?O]ӳgD\z|b)U={O4!T8S޳JO;@^bd=ti~š,XF?MԔtj6<,)8Ha?~c)9LF !8s\2+̾V2ua=nj+ 솬FsYMqV~ O@.xjw HG{P)%pa)rqbXNxXEDTH2R-!z04taK3urh@Gڸ~$!qd8zèFΈzzwzYcvP\;+Ŋq*lH1-`$*h"sI0e†u򪖧+ fAQCRRWV6ڰUH~ 1"j&9J`L4RF:ɽ͐w 0RQp*7UWkvU-kG6 fSr ?%34޹[X=ƋayRu<^A{d)ªӌ8{T(S%n%I SP?V34wI˻'Q(L UubZ:0i-0 \1~ay(;9X;/`W棱ڋ8%ٗx р<׽p `{0Wf||Z_Ya|L?%ztALofE'2@xOy*V2ϒtw _m"ev%jh#I4aYJJGJ-}zt ,fn&Bgec#sfY|m)Qgol}$0I>ǺݍXX3LXxG\A4Ԓ4tH2*v`'`SAw=Fq1`:N+93^IH9ڣNvP}.~=/kܧI>_ w݊ E!dFM`gvy2濑uc76qa>tX,ݗY*;bA+­p &4  @ f:J p+BPQr$h f:Js<%v&9<FI O b {pK6)>л_'.f.MBQmKJQ#T݃ӠC_>bB3+K!_/)M>f3'3tZ<<>M*&{S :3tpUC MɺM=?MUoJ}Ru!aRx%w[Na K#2ڎj[Q6⮹\I^9(= <`cot֞‹2L`T1⯤9J|F#q&m鹺xa<2X܂.)_Փݤ^cpT|}]BAQ| ƞݔAW@Naa$]"t2iz% duo 5wy!)&^W5a\3ճ$_T G3J_cVQRݲME2d~S_Jj0 >¹;'D3I~w^9Ts)XO'"mr[K Q6hTN'fd$\,(87-8Hڏw.I)לØ N:r?CS #K'2BV1*D~1W {A!g4DDaӯ{ Ki*gd|NRC?=Sb&'1;c؝mY, _ vtϟ=^P1IdEIT,a6T^ V<;0j˘Dű4ɔepRHZLDM/'Hԥ8{d=cT)"D)T8"X՘Sf "G(_SjęJu8o[b BU .m2"eYSKavSc) r$ZOi.mF/|iWʄIeb#Bu hwNXјS p)hb=a a|xiL+ LP`9g ,@u:r, 99v4pZX+bOkFxs+ݛJ7f7ԲN8 #B:9 n#&N,ߠ虖뙍m%J'˔vz?)<ΟDc_y@yEF lF[` DNEEF)ldD1%Cq(bPCkOq|=rr[ oxr5}o#BD2"(H%%19J*28U74! ,|U Ɗ%m@ kOC)x_)dUFJWGPZZŷj4&c0]J9iiLKJĘf'fP0!PR|8¥KwtĞ&rd%thy0&Wze|ENXi)p[tȹ%5U}X pj'рsvxD] cqsC:I= Y{}JŎLxŁl@إBXT1d-^E;fHoF8+1*ő]] sjcD01bZb #rv1]_]A pص#X5.J]]X_]LZ uv1jbK!2Rǵ2{!h^X3`* cy) (;##/~[.~[.zT\]L%ZX1}vjb%Jk)kǧ)_'wdl*6IpKI/J)LcoHlL0եTC:C߻ +d˰BYVzTь9khUp*x,5f.EĔi;B-?cCx#d6iAbc# ,8]v[j@suhN^{ט?"/NQ(bK:&0y|, mCvn.ҽZ\Տ ? ~k̠?$,!(%PB^=la{[z_v:;w&!Lz;_֛C#Fa/\ Vu'_} gO庂wA~yLO=ةRcA龩K*{}WV闟~(صƝƢeo3[ {l+5yiyv?Y iIU{fz͘wwl?:}z?ZU`s}yrP52S;>s5E\5$A$49iv.z12=v|Sώ}0x=i}Vq"cMQ\8cldPZwM^LjcUN/1}<*_P+i+5靳R߃l5R7"9v,%7\k`DF^&mdV}B#cKtUI׶Adm{*ƚ!|Lw  E)' ʙٗʳaw+=R"W$ql0Asutgʚ۸_aS¾JN*u}bW:`S"s+64")i8n*KAF1vޢ.G1lxW T+Z؟~Ɗ嶊uNHҒ/oN5m5: TsdCN|9ՂQ|mvs%MN9ՊJ6z[SnNZztcAoOiK$}R O92!7mٔoCK}&F jܾӶ7r2a06M؆dta"+C=it)=;Κ OO>k/rYE)hͽ|5,PXo΍7󥛏\a5jV?_dAv|mi@v@wO^f[ ,!yy}rީ+4vȨwkׯh~J~`D|vd:,˦t zl֓P, j5j Kq =X7u0nA1ߦ: i=45WLoiQ)Gǘ% ZyAP,5A)F4KrЁjsHCDأkET_,SM D4S$njI'W ~ܬh+:}sYw<%LY_ f,k; i'[|辬k.=sQLK?r2v$䙋hL5n(yb4 MӲ n,'9.WNQkB%Y7AXD jNYtUnmZg.)2Umz Ws[*uD'uM|O?7B%5&Y1m37I2%]]c"ƤbPGtRd_kxҝYֵnmH3$EX7F<떊AIc붿Spwfݒ'Z׺!!\DdV;ԶMVѺbPGtRĺ婢ʵ ۶nkڐg.)29'V r;16nRxZmm<Ѻ֭ y"$SM/255+M`.bPGtRĺSֵnmH3"uV+ :D j=$Q]ޮ(<wKh$䙋hLᦗڻUV~i"16Lw!E !\DdrvPOP(tA>]Ц&q`jr%zDoW$_BƕrlmA4?zUF2V \ Dux-jArǡR5[fUM|*Z\ߓ{0?Ud$ׂ YZ2r@1 PM9cnU@ .s MT]Q1s1L9cnA(r̄͟9ǜsmhgrTs9ܪ&#B5.sMM`&3ㄜc~9f昙.j1?3_Sr9Κ3,9ǜsmj@Z_Yטs9ܪ&Ɏ/,T9ǜsmjxaGcRs1 aYAs9ܦ&(5>cnU4Ar̚is9ܪ&>sZQs9ܪ&(!/Ǭj_3 y2oL>u_h<۩w,~`㽲"qNd&"N%qTdN =&)e ~ _bËr$rxp6{>/\blbLaa8|-g1k3$""~>N2^{ھG]t;llY¦$+s_vHNWio~&**ee ܷhP4󏷯/~*]ɧx  f@s%e BBNAk[xaD0I-( (%֪@FoyyA~2Lk.Xs''hfWA(RI)@z^ ^H # 'HzWM 1rFAX 99뎽qD/9c$\`IKǷz1_[Lq:6Myoxqb`鎛`(!|ԡw%'ya\o^GVDQjHM <бpcyQf08 jY:I (|X< .n",B07`i*T00-% sq@^PkRːyh MA/omiHܼ;V=o2E2 |Fm> nq<ͮ R$2?^ksy#k/1 j+\q[rs?:"z3@/  bA"MCY{ó[r# ~M\j:c<+q/N)%Ap'@1/$׺RR}!jD2/7@ &dP j!T=fJAx] ./:C gLahИ[& b)L'>U`҆- SEC,1(FGaIb/AALbGQOb:G%U_g? =Vwj sWaZ'/y;ǩ__,s0п(gT$URmȣ,}?y f5~̕2 چ’:^{ Ku~^%$̦~<9{4cl͕jkpzy'ez3̟Sn3d `YDVFAέι8zJU\Cӿ% /?s..{EUnk)cعXBxPUL*- z6M|~HG Nܠp2;wXT+ʮ߆i&ǝma=:QMZ=͗ _iA'˂r4xQpDÏU!X bdBsM^iQH+i@Q+E1 gp;\arQF"  VpLLl1KZ3 ,kFmr *X':OX aSNp"@[VQ2Cq\y RF9nGQ/NKٴ}bjHe¢t:W :a<,nmLY՞ԕ*:]Rx+,(S_imH3$f`R1#:clbxOn9Jh]ֆ~JTosI)#V2aByh;3|>,bFo产]烙 =7-l}6'S*jL-ߞYfҬ ^# =QNںGy<W~+-+-@^~Ypj{vߞq-LAнޔ&Y3k#GEe`9iΡO϶ƲeKnNjKagUI*f13+-%h[VYd|A2"MʞM(C}U׏RK=/jd+4%r|ؿK_L\C\Mxx^l,bq^] *92reFgo )gο%ĉIRa,UG-ByC<xzlpVFk4?!:%D P4nV(T!B?|E Ӭa9Hh3\F_ V;$CһH(Bo[+R}>zcrUΉnQ듿x:ޜ|5M&zf6@I $) hх`󶟻]PYMLA^f̷uS 'n$ 2AARB>t6m'P1n)n٘iP| G~|s|o&~ujR:!NjK*z IæW:iFϽ-[H1`SlGRu "r[WPN[kߛu<-3hܲ*j(Vr,|*eP/*2˨uXs,l5mL}՘v<GmYm{8emaҧNZÝKZ1-P+9?O[Ŭ?ENOݕR~+ɍ/^_g*.pV(YQ2u~%kKCWLJhļJ^$Q+4!X%10zoĴR'iqCk_JǛo޶SbjfQu&nw<̓K=SVc2KU"qR9EDzB z3fKfD{%Z1i=E@zvU[&-$ Fy%^8Yx `ey1Jɂֽ2e8+pVᬯ ;-͸E)}Ҏk:p4HdE _;ցZ?x#jk5yPcEkZ޹fP`S4Yv @FB484h!L6=hI\8m3w.$OR%A=q8`L f[wԲتTokIMޣ 2reDjlTQ*I%gsg1qZlo/6]>0oX'qgvٻRn-$cћU$פ[WZ~A.VgGd^E-f8uAV O‘YFR^D䥚VeJl:!:ז*)ɹ5UaeWJy?؃ xf̷T"mdZ8@^l[a;U ,Tz.gvPl*Y:U}xHNH5<W7>o'f@dcuAt@MѭRc#Ԙ&,q T,I -#+GkLqUsIg?"iHZ~ M~[1i.3Fz'Uw|oSd\t3 dԯX9n]- (7V-Tp#4Z"䰿cx55je)[]]MՇ)eB׌J_եl‡1R)FT%[UטlQ̕޼Zı,b Ϲez53fS!`lJ]fpUqtجlNjB3^+Y%{Ge ⻏ L,;ppm-Ф'(^8 Pi}=Q.^n1ʼ,iXtAbDFVwHt:Ybލس5Ol+NV 3*(2MbkƮt]JZޏuч)ES;8hr#}}sE)Ŝ[*y!a޴Qk6缯mA$t"Q2CMsN2f#^G=؁4cfqVsЂ2Үc)uYI<9/Fzy/¦KFILiܟ?}>LD>$Lb>svDq5ױ'vIJk* gWDD1{jf<(}7쇟3`P)#CʘЀҼ{x{Sʞm8'֝Δ!rk=SF麭MHܝ~Uz #c<;K)8DZV 9 `⻊>6 䕯z1&smI3!Ww;(OCRq͜^΂S&8Қpք/D =3Ի yfQT+N`L;5$miz2KqXMέ6p~CT N9~rT.ʥRt֯\O *r9 SeA6DH'^T@@M/~ҾyZ/<3^!4.Y8@J "R-^6rA NZYyW S#>1`қG$5aJQ[&11gl3ټvb 9:3M̻S!X5&55&L!*{hͩh XN ctɡ H2<ѹ8¢6"OJ+&5IX5(;AsaʛIʉSL2CQ| @^&b4w8yAqdY^ۿ'o(Չ xg%Ny"4>f }DٻFndWHd8 d쾜 {,y%9(vKjlj)=2cK}źuIppЌH[dZJ%8yEsG4,,r &}OBH8sEҔ HSb3 rMfYVViQ&5s$GSexQ+48.))ڗg>/ -A#8L5S(QXΩ*Q cTXDOsɴ+e^*$Ip輔& zfL&jk+Gr܆GJ2d||uxf﷮/KL ih:uxA uMihB+T<} ]ɖ1Cuqroq}oA'4?~bq:̾|DJo vUWP;38>z (&pTqZ11ɠjV';d8)ml(:h9+y*K+hKrT%dNrJp 8?Lp( 4Znʜ™.<ל%)3G(QR Q$j>F}Hyh}K՝q3T])}??q4oG4;ցG$B)b..ʄ " 0Fw\^ ]AP[#ptRF]f]жfL )2\h`y=Pw~]!F3x4j~ݿV6CXL|4u)W*~3 #h͐Ŧe> kj F_ĽcG>-?tauV1"~x\f6 =[zzV=xlyǣO/#hHdHJH G1sTiQI 9{㵞 NRI~P1J&Cy)M dQ)-Ln,@jp( I9K{hFJUwwc쭖L+*`g|,$'uT~?7a[W2x7,xgЛ?pg^ ͯ'y;% ̄|HsjHڌbF)c2L E)MHiI=6$ԇ z.S?hX󠡻B&bG%x ]Lr.AjcT$FjJ*<#hӑ~HJn?U{^lwo`e,8kw .e8]%dk+xRA_?).^O'u͉U{q {{QJ>٢\NQtw!XAoKvv?R#< -+pUs)jBN/zSpT5^V\jA\vY:M.$t\;`I%w*hJOa]Iӈ(s,dȨRo߫,h(Ep|0P6ҝntr2(0}⁚%2_LhnQ*ZWʳcDtu)zvwMft ¼FZ"KPHh9ᙑeڽ"5ԮۧeRE!ݞ AgڇFZ+YT` 4/!:Z`Bh(TJF˥um`,i4ф馪;XT 9ႃPɤBѰ7'cəV`eLKӱH"A7k/SJ wE̋;k'/Gq2-x -EGZc'QBEw^]T+l-dĨ*͗)uh 2geT)o zZ6zT˓YQ%"Tv]Pg裋L$"9dIj<)&*9lɹz'gW/&t<(k/NȠkezxF;:uyp\u*qa%~:Hd1Z&N~/~A,bh硯.3m] s !F8C, #b `Itg`  HҶ:k!-< >C@/n\.Z&Ogg7P(Gߧȿ75flM2^Nq?&EǤDWǿI'^uzxmW#<),Mʌ (dLH *#\H0涠PtF[oG dr2+&/gd:ѷϫtV(Q4# V9G;[du7dj1iS"a>B)_]?Z1gJ0\^={W|d?'Y(f~=M6PrGmo2`6Ơp | Zd7ŋhB@P?g˻IJDT0$\)cB|!Q^]?$KTD ,a9ϹѸvQ$,IX1}fؽp݈竷Ȗ˩9XJE@k<hQA;JEdJ-6Hx!P`.ЗaVIZ)GI ocF'gHUd-v.kaTUT٢٬6h#\);Zµ&vvtBSu]avZ$ cZL2K DhmFuREyO},GE 5%|˃v O\T )?\@'j68q /\ d(?^\X-L/^:r7'd@Z&W.R?J޵"/Pr`?!Y!$=m2FC|)tJ!(E!ҺƃQtᶈ֋iVpQgE>FSPsb"2DrqѴW{ֳCt`#]a؛ڦk:cW5xHT0F QР{u߳fc7_!Zh#|Q!'^uzxmWh @TJlif YYKaxfpuޥ4fT4x6BgޅhC' *D:L)){^Esft%V(==tϭ+sJVKT=x )GYMt+4+}Y4@⯢ޏ/plZ 7(dwz9 yH,Vh]5Y/]~2,мyHK'Ĺןo.ܒO۲G;8 ٴp 6-~?w|#JS>RiHVC;XqĽhKuK?=Gꄊ%Ks # ];iy7mJ{I}[œ9@LЊĶɺ". G#^D>~KB Yd@pmBK7G;4hŇY}2MQ+v2[ {7{诳ӏG_&wGvfLG|2hab#׆5~;G~!L#:@ng33o΁c5Gڴs izY$A|H`-썍 `!2+h>` =`hoN)<OS+?]5n𴺚)8XcHX'#ǭPt .Osk ÈCjv՟[tM<\s+JB Bnj$6stn:+ODk2ݕOZ dV 'L?shgc6K_L.$dlAȋc)a$b0+NbӚ#Iͼq#D!DТ,mZPT(Rfu6)R:/Lu&$ |"L((ɡ-G@H?"aR~CHjqĽ4hNoЗ;9, G d D6hhnΆ U1}(/ggצCh}0[݆d*8O=k > 'ۊ?\Vfg%7h|=b'^'Xx:Xy0$6&`ui TeEi!&ZM)D}DCrS\\?|@D57 ~wӛ_[*GÚjIDzFJ(GW揼D &Ow&&ON^ڼ`'܌S HO6~4n2X7Bte1nF!)V]ۺN$UFl0n&.$I'GSet_h={1Yr ۲6ȶTt[zt9 nh~uxK,n$FyQ "\'S4&SUs`D_@o3mQ($K6mf[X{jK >6TV])7q?BeUNܰ>RZ)VL# nipN'˒^5Iz]`>NCH 9fYmͦ(P:&dqp']F֪$7e)dtYƓѺh@L ZE9*<< B̧!2 :cEFI"Jba~M }gCӱAma&Bv6hsJrarK/˹gG+EƫŨfWU' Wzx3LǒQ큮$)1VZ$84wL.!AD4G= J"7x|ZU"7|Lr>zpwh85 I?3U| ߎ1V*h[cz,kY=TkvRmNibI xM{f'eATvƈ1ȂXE =h"iYy^4xX]2҉e05[ƛsɆTY-j<]Ew˵rc88*zawX=R)5b)eEGnY5:Z,<ŞͿ䙠a≈գ}Ѕ)JՍupd>FA6Ompnq (D.i^?UNg}P٪t㼌E>O?yK,gWz,ۙZVꍹ+Ȧ2-j!3D僥o/^ՒM/mZӼ>{ɚ|v ?yLs%6X\{zD9gpKBR)Il3xx[ bPU,~J`0Uׁ져iQGeU_V{O>ѴtMSqsU߾]x'E<|n+[|T܎j" u‹2-yk^HqvO,pRk諎7qN>خ>Ӣ(ލ>->T8=f#XOPn3%#l/JO=ēBny5rES%??}~ui?} iR5NZ$McHȶq}ϊ?^~|^LaɆdGnU|_2f1S0EkϫzWϧÏ TfS7GvҋZq8sb;L/+'L?[WGaN?,*o~v,5jd 8V+Q.˽&esgLB)?t0 ŞT#&}1(G$ՊqWrG7KsXD@ >RQza})DsW1fD+|LpG7 qĵO>*5fL"FVN*HQ@8S w6%9f%;Odl幠X֧wھ#(rT =PM UuCrA,en ]WCH'U54܈ WDH(ͅ DD,"VX` ~}AHv\S ,Gϙ:"P8 P$= hc$;.˜aBPT)yX #eb 2GptjyUFR ^`%A4Cl9/xk~>ÚjID1h!Kbpg/y1AK'rBF (fXQp" NhFZ@KIS_Th8/J ~]I7x9z4I Km{S.E^P<1w{_:Ra1#7dE &1_Joty"vdyd.amSMJN6ZxSt_3i7?90CO|HqjӇNWcp\UׄޤdL⸉>%Sr{TC:vҒ)>^^P/)|#j$i<Jx^ 9[䱅\K7CЪ>z#>T`!N)`-toHs+JtD UPM]%|y]Y^CB1^]gw [:S~ #@sV9wh.B̰;;63wck[W; nG{QpƺĖNNP7MIȩcGҙD_occTT9( E%?CS!`/70KF&'M oVh>! RvR3M(")xvqz~}Az P ܔS?3m=?cڵՒ;by;d;x m3O1ϋnS$yl=^Ԏ=>j)30JW>3[oy} im)<u~7h P:5mL[1mx23n8H΅ӸS(\P?YxإQ e?6󷾪֏o b}Bh# K׌ֱ nx5*u3ʥsZ״ЏqJ)N& xqS`Q4M#[)l&5 +M5ⵠ \#nnT0.kB7Yz5$[^p\80Tu`S8qS͜$leaSh F79D} m Fn\#=8-lq SF C6{Ceceq4HF,iaTA4hty [@_6ԥtI = V.$FOL7ۀM]t7K]Iٔ3q2f;$\nZ]Jy1j*e:փ LnC(;sk > '?\-HKn*417g"zŲ;F\uɮCcDyX@=?d2P(qƞtM}&5&;>h2b}6e>˜g+hE ͝4Hډ0]mM|مd^,2{t:.4pΞ]q*S./֪d1-Xj:ܦB[2ߚ/4y0!xdț7h1 [.G[-5q菤b0#%88Q_8 9%7 .Zb5ql|X-80.-^) 9%V00'%$7k8-iO_q;#(Ӫa/.y 'E;ƓwZ4 GQ|_@!<c7ho wDcUJmICLIWʾ=WusUdr}MlKwei}N~nI2>\HNS;Ԏ>[I[K~E}^etSB-BvH&kfn^wIN o֏y,y /}ޠ6N8sn1'6!~,/h>Nǜٻ涑$W|@ܒz;v?,nSݚ/ NidGeUeVeeRM.pY)Ð&plD"TO Ӎu@(C'Slϧƙi>.Ceb^H r½ :z5t Q`߯#w{% 3&EuJxIOy_~D<͗| 4H nB_hds9VC?1( )8R<1=oDbo|A}t|}Gاw-+_ ~a?wlH[ov bUv L]dΞ+rq)1WZ $*S#8YIal+0=ޙd=xͰJ>?qlER*4m)ʇ:фI/D>'kj,(VApXybZw*TTsK%&>n)rg3vBrRuÕ?_|)Ue ٻ.|Y~Y1y}5 N*)* u&pyF(v/8y_F"1C}Ù8tٙy$ybPfi+WS}[&A?3/;݇~DS2n4e+DVuEIH˚.%T+yxfRH\pI0-@s9eDUPxU+[IRub@ g!7M[쎏~ov-uiqH2%L^D I [it>n9eX%gBwJDK9].l#*J+[ R2X}ieϕF8) DH pLnjȆH̪*cH\ cɪ^0y0#(e,$paRCĈE1f44,:APk9@2}Lu1d̂ TN&Ao]veId9=ҙ&d,b{QՀH.;jt4 gBeT0E 1KTF~x3ELߝ,/b)0Z:sw6^l1-*ϏL:_Db |OL1ڪn1rϛOVEeB7GR*ud|ق ;s)ia"bO0A9;Ć]ha:Q! yN6ƛdW󴝐qOU|M3ꇦ\#6"J<Ԃ# ZDVR{)9iZLgL{BlU Ϟg݄gkLyR(x`SK(k#U0EE|[w[(io.@Yu_!}fW(H./^#M{\㯐**Zgh[}r %{[I4eFXEn_ZگGIx[ǜe_!WwL<#Dsqxo<cwmo,`cjsu<ۡ ɉrō%.ys*Gz8hd<]:Hy.WY vJ2DT͌BuU ?։QkQWV#.eb^u8E녏R|VΑus6#T:U"\vJˎC5h. ij.rDk}.+lexTV/}bH7GeYJ^jeߴB8z ()]87P>(5o6 ?S0I:,4abcp3i"%JEBpPF ccz7˙gaR IBQ`$5M DDR*䔑b!) :VERSǓHi2 `d]74}L*70^H{v:L{u[$d0N5T&f dʥ|}w{CƿGw>W?Hw}}zRя}μ G69֛]‚go1k/MHrobd1 EԊF"Sh:PL`<ޯ}rK&I6o_&fXGLdzlg Ŀ*ݳv~LyM,'l5D7@w}pO$-Xhg268 4rMTֵ3][@WMnw`q*,m(vtj Nj3H]D_bQ/F,tbJjN';Ejcncæ 5: 8SMMG_trIj;4&B]%s6yeA|k˂.~6yFa8]Qxx=(G.o#rA*AMjqD\RHk*),>v[%֭9p'ĭ.j[3 Z]I9Jޅӷ'A+=j}[f1.b]d@V:B cf )0jh6Ktn3 t(ݓ RnRX?/1FjF%jzv1i^8ũZbOuM"i`ƱRjM\lC | 'ZrhJ6@PZ4X+NٵUwRӬ]σ*)Xxk!AJBH' ;6M M0_̼x:y'ٽP˄˜ӧ0Q&+&~ 3 d1x9,Li&Ϟ{T|PA{x4# g)x}@6#B21ܺozerX&ȶWvrd/?ړyY~hS8pvpt?DN@^"ԅ%3p}fdSœ7>=$ryKIթoX{Yvn&U  >9 s;D@Gp3A5^?20ӎL~Wycp8gO&W@x0+W/Cl%,{SM9tɿD}%zf4p*7+6|x1sUL2 YbfIVsdA{+I^f" i}o4V}đ%6^O^ y:43&rO.ung"/ S L0M)XdkN~{{xB_aֲcĮ> =,ݮX>-|Mua5VrʲY]L>n@jQG/Ak̥F{y(3]ؓs &6D$D;th6[<$#R nܒ?Ȭο7$ dƇ,s* j8O%kZ1}m׹<U_~*(&HD#FPl, Gg\P $ <uE̹8<$T%oxڈYD,BHŘ#Xbo!(/5xUpxw~C_yڍ9K2W.3= z5& *ݮ.gʚ8_amw:t,D8<d4[C0`̻4 ae n% )I0~}57A㈕nwд^һ BtƸfleʚD)-6=y8FHɰn7(niP0Sc )!nljL4>r狑V{^ύ{2kt|ZNhq!:MyL $ $(}@kFK%l:AL@r}􁜠8n#2M$5sHObGiؽ! D|skS XMhGR{H=]>8D%>} C. ?O@x9 Hq" &Du:ހ vE1th#]n,%`h`pQ39r=ZӥQ֓l+I#GСC5h4CєNoi)MtB9Ir-h(<"9OW&3 $sNHҎHi@'sY<*cQ?\J+9Oo faU* -"B؆<5)ZxE"̮A"?TcXYJ:/ԍt\ VbzrEbPTT'Sc)rV H0KL@q6hR;Fwwweɾhag RIBL;HpACk!;h1<Dޙ4xMPcD$hBS<7ɻO׿1s! e jcPLg1 Y|]F{bpR@lW  0Xde&QZ8<#F6I4V(tߝMK]y:`ojQ څ"g7gyڟ^\@ D"?/ Fm){o?]4U;K !,alhYx\AIx%ˆ+C^NH)vr\Ba0٢b$j s:%f@E29u#7bwOrz[{.Lo`2E&E{ }ԄY5KN/nX^o -W>HibMUgs]hsr),:+9d ead!`sV)'uvīJx)İxyF>唅Oxw3nRdpe$A@kRXc=PؓG WǍB~7)'W$,]_їxͭGNBkgpuZMT&y&Wdrr7J&}PhJOgCV ձ{$kewL-m 8H -DbJBɱ#-gCRh٩j%dWH)RsTHa.ٓ.,ߍ'7wWD/3!8zQ"l&B čN׏WD (ܑuulQPmCu*u.Q;"svU"bk4ڔ_{ǘkv7yR_}b Zқ,sIr5} p=*ח߱!8#|4|REB؅ Ǯo4 mN42U֛BPֽ+(y ˖$U`B}W %ÕR۔-paDIHCA.-(F7IG;Pbka= zyqu߷qqw0=##WLieϓG~~_{^7+wus&}ђjk݇mtWȗ~(?N#Y?O ̮5oOGPߒ3흋ؿF"w/~ֆ['eF IhmʟqZh=$Զ5-cl cbof-Z-uG;vzL.N;'zEy2u,/ʱn,'{uP7s=OWݔX&BzcPNf9썾Ϙϼ}<]- 7G O <6y8ݟg^M`RpsKt31MwIfAfN 4(x]ir~6r%SSZ7֭/5]᢯V][ɗm y"KG͹n}1b[Ewm_면nCH+%>0r6tfLϳ/s{}r5~?{]6 <)gGSX2T5#.L$ލZiW){]l{Js^ftkv=Q?/dRΣDk+`rs<_,ӯ'ӳw3׃ʖvCe9S\= ,ig=ܟ` -$֞8MjG+J|aʌc\ jIlppRdb]FmDI4734.d ϸσ\a+!_Q'UVu*~ Axز1y2GHVl7VT|]w\NR)s,^+Z`_nZuwS7 !ڪF*7¡}dHҡƁ2 )X[D4Yd@2*Eݜyv_SvZno(ҬR*`,q[qc%&<,M2:Fqe޺ +(9R5?|&y"h}BxmTI# _ vao(e2Q@Bʾ=tPi{nkN7CjHTt7QF5+IL$C0lP]_iPRj[!&]jl3O(\LjHR j#y` 9G'k͘P`gN1vݘb)R̨*cFU^0N:UUۉmGignp|ou>iM;aE#&۾*XSu% Mt;/G{uHK-.ZTP,yko~_WZļ NK2>lKreQk#̒Um7ZKm \}!:7(~S-#~0\!vu!UnjOiOv)}{rC'usp| ݩ;]'-:*iֽ7·\s̺1N㗐+vp/5>˅i})]8+/!!\DT~ֱz [_ jD3XFU ޮ[|J6r%S<qZ7GQ,/5]!WzgVZ!$䕋2%rz{=V8!w?BeB#;ٲa;jЅJZkٰڠAӒ'k~玊9WNI-T>Uyf @WG(tԫTF4g98\ޝ~joW&[*_ku˴읈5lQÖe:2!ҝ5,1,yQG؃ɐ$֝ŧM1;4( Wz_wrfI~A!%lO l$3`ZpJ7BJuyWWT'9W%,NeF6<㥰?{Gnd"/g<%Hl`n[l)%A!{FrOkfD6NĶZŪEV4WX>fjg xޮW b مF2UW$OFum{(7FhLY5 I6( S jc2R^:w"1'VNZJEJ ŨbV DQi 傗ŪEU ΕWҴJ(mlޖ!V*aH-%\PPY*JUZ9+C=N)*ЄУDfN60zɑAm XPMQp W~+;/^c_|V0E/U4۟:۠7|ٟbxbʳLJLJ~p{q~T./\(V%)ك=|_b"zv틛jao_(Y4Bo]eZj^Ӏ[U4*\!.uv+KFZ$ehZYPʋo}ZՍ0owCf^dqM``(SR qp)5<xQ^•f P'7+OdSUZWB 0NPl82e/02zAh40Ő(FH%ZY*zznyST+MУ /0\L)VuqLÃQXGCQ)VOnbtD`QFjYBIJow&\T @µLa;>3Nա>H.D åV%QxOMS  {U S&'igfH&XWxSދ IȾlsԌ _.Rs"&A9>oPN6mOM96DRNu7,L1=wFpD~#Ż6d8B[hޭ qfa8_x7M)x\@')m(O-6!һu7&[**V oq5bް]< T%0%0JQ7t[l }gkN7Pΐȭ7Cb#P]GN[o9ai76g2%E~+]RUR{wV"9S!j[+_>뱿;_Ůq`[Y6ӽ杈z77|!Bh $o%Dz~I#UwzIm\7k PF9;CI0[P%^T*It\`(%Rzw!е S:?fҤ4,YYI`܆4Jax«_ʒy *L)DYmB#=l{=v)=6F䟰>M /"X;f"d'*gO @O=:$ɦE߹S%fo, }Y$[1Z%_۳(*eQ S5J<|a̿K( b%ZJ )+m4U:Œ"^yOp&S9uͨ5VRFuǎc*%=UdV>(3dE :u\it'*6+g xz%qHڥN#>,JpGdQ/#sEѲ(z؞EIW>,JFrzY+3E鬰FZ`YI,(@ -qoRŁE1Ԍ]v[笣O_\_U]"BCUׯӊFjF+#luDcsEF yCxi1֦FaF9d/"ceOWdFʀFr6c1wE&c`?DTJ Qwwb7UÊw?y76--݋7>.>YKR2ܟm7#Zm 3"`ghpHa60APHq5$Ɔz܇'ZPj9B e\NQh'~5c3>M8bn,z[Nik3)xv( Ɖ[i$<3o~]]M6B:b<B<6|sy5\+O0$_D=]v$b cN(KjK[bZq__K4% [@(gJs),$:pn,&~º%cim(P!7?E $ 煭8`H1G\"CmokĻPr/xW*{),!n߉d&ɥm]s!7ۯ!Z_(ι56dWwU< ΗX%Hi^ݖ(~x~ F;8@ӷ9kD:|`*V~" ^w8s*T1[#nxmiͧ7md64I\jF7Ҕ{-]93ֳWMHo7&7e(>9\bA{jn^4ۓ|xpv5oe}]9rυm$r.N,NU2|pGb!_3db3tI|Yn QzcN#uH(&I͑_}D6SDejlpx@Y盻> #|24;aۻv/W.Qz]?53_>B?:/ wAǞ5~;"ë`w#'׊EŒ9Um!6 fɑ7hxVK-6q(e|(ђLlK BGRq(T$/'e[jskF1EJQr'}R_֔GRvAT JsL tԗm5kvQz(HZ h9ιx<^WzT>}_zoma%R9d dƍ?Hϥq1h8u|^䕜eUk/.sբ F.8az D/\ Kuʄ/l-RDBSukOe8s)pF=3v<=(QXf@K싧m^Y큤91/#fZO7"flPs~E|0S %%#*P@$q,DgGu>[Ÿ@vE*,%LӊۢBA Q4e'îq*@ŸDzy.Jw]st =J`ӖBYVZ n6V  -SRq%9y[Y;o`~*~S}8rL+8[WW o?GV-@YKLQOVT;JAJ8;WT242JRԽ6T#]q=nߛ|o0b:dKJbxѻ‡ɤ+0T3[ -ӕ**nJjf8%%)rȼO6Hn#K0NApku,qhտFwVj7QR S Jw|]8JSK\iG8$h'V~&}J[^FiЕ5J#1h1FBHOQؑkG ,O=xX>+*컢3vM$Wof՛`+~4ʹLԌǝXc K .M Ju% h::Ce.cpH5!0Nt]nx"88G(`#OQH?Ŕ `\ՇTߞXT4Q}#~:Z 5W%+~V.F'\iU22r_;7܊mS,|U x4<mP)ް$*$򠐱%R2݈+a$R*c ;/I^)/_*|!՜X}E8]G4ςOzw;ODͨ[~kII}ٖ$yƎ;MR&@)bJWR Q/ ^hx. :nR_LGҸkJ)ġZ# PjFH̏Rq(e}7JC)ry?&JCi- g1J1֝@Ip6-0KqX3&JR+/9JC)c|@)#q(3_qCJy &Rq( RSBK/'Rܗj_:e8j u<^ %~/R st(R8N PZKQz(DPJ㢧Ԝ{q4.zOI4.zZIW|2w;nbދJc#+Q*#u0SXedԚ(=r1:炽c4&DQ$@sP ,RdHFB-J2J }+e IC.xsKo0g6u]Y|2?Y_˶BoAD/pQN{l7 dTe4JDM&2Ic{2I:yvűGԀv,&@ zjېE@q}~.@V"j&)* 3b7=W{t*o( :`Fh'Ť)Fe qDW l׍m7 Vrjm-r-J㾷io[&(q^"!Qb?@"2Qبa),%5R 8f٪[?BՐ nW^৿.7.w`"?D~h"rwwu(R-p %-(PRwD-HҮ#4rgįZlvuC3N<<noo=|*m  eq?;ybNr5ⴉF.4{8m ǯW~ Y$+ oH9Ǡ'=2E/Ou΄BB)-\vah)Qٗoݱ UJʻ^߿i7nA<W?w K={кU zKvY bU}'?AB",ݹ,T"A1 lRE=>֗ 4B}6ngb4WyU9u Kx](ʗˆu?_ S`&~faz/̐@*8CmkH)(8M<ՂX0z zK`%rXd%% Kkc by 8t 󾔰))}©ňPN (xR\ ۯuxꥠCdmc)'~) 'qv_ubӖbəz(1/l8R k xrg`yڃRDmX:<4Z3JĆKz9Cロqg5.Fswn|k?! }st>):0`G뾺EͣkGKAo|yq|Oi=zY}ahRmݶcz􋽮;Y~ۓKS?Pn,i\ Yct?t<SDlY۪ 89BMlpQ0.75-7yٷB1BΞMD~;]P ^nڶ] NV;_qiEvWXM=.}H+p\ܗEq noCڬD[ַ ~ha\4hZylE}W$b~ɇ|l#_*EQB/L;0g㬈 zz~y$dՎZ58BHc`+ĕ+))8RA@rW`j5<(}׆E{(t$Z,RpE"e\v&њP8Z]:n Tp[  Bhh"QZT+3iEyJ3B&&/& ,3Kl 241 -;G"A;:S+.MD52E1)dXqUT3lEד@]O6@lc׵klcƦ>]:tbu@ r3YxȋBrYCȾiYCOQjN58NQS:E3!Shxi^ܬK\>Iݴg}gd }烔}C5>iWS35,i Owhv/w`^U '}G|ۂ;ftNԬ.ZwjOMD9;IEEP-h~Xé٠Z~V'% ?WˣҧYFRzRJhhYIVdѕ2pNJ3lKAɢ+2g.AR7+H}hb/Gy_wxR$Ϋ7߹yf]KS`dM.`Q,)B؉%&KDw[{K5Lc"p+$fliVy$f Vf`GÖQL)ES`IzmKdP ꫉)R ~g<뷋'V>`Pc6p&' t;|\b[NOTk>{m̵ȅ%{AoQBwx0>3T{|b`3{-1 0`VXl/QWXS*"a`ji-0,Πb#ykD^\W^㭴?ڋgQțCm9ih[=B.5;rjgGk ' z1wbwy\Kdװw~..>F̸cx/.nSjk'n5rs8J[P|doM}ms:_X(,edwYWġ2` =Makv1z|~Dh@Z6 ) ؅{A`JxkWi#ۮA@}WiMĐaӡ051d;6Cf} լTᄏl|ǂѷػSvvZ1EۗuUl#J%7}z5.YRWmnc1!;F6-Sy&6P`&$*{nǺSoj7QfԢv {u[֗&nڭ E4J0=z~CQ1hX GQGO[|HvkBBq]hq"6n"Kxqz5tKbO';71OW݋/fz1$ˡ/st${m$pA٘>(<=};(<Dƥ6)o6\\@ohqRX=ZctrR/QP; \S]nW~b stz˺{8nzD3/V \,]5P0utG>6YRbSEFDKvBKa_w }_nHgz> $[ n$ty 67bPCX<|?s?.S?)0S`OmFnnP,HU-)/ȺmUqFs'$lajY9*l&kRUF_ 7_7u}:>fNf$ȯkVˠQ5Oݪylj=9+bR$ɶDfWHMV˺ʴ*yB.E&`EU*lnDȿf:~r4Gdۼ"UT(rV_= C @"0tĺ4Y*G.eҮ seYL)m2e+i2ʡœ1d~t% cYDz8X; @GU.=eYߥEfҕFk0*8jd+" nepGsZ[ ӦQ~=jHˏm7+ uxQuHPn h=xg ٘L]zδ`AZ{92N+r#AؠKΐ Lց)*PcZUTצAH1*2C*-P7~aˇMФ'<>*[!C#M;^0ZT0ɞoTf5_xҟ.lbn{m- M?^Xo$}ˉjsq _ -wofܝWzi t"Ӝ? ~p^yw^3h NZ | V;tѮ!/W&OkޟU˕))&?Gt[*>#&qu}Jr薞?Čn}0S4SJO~ŹL[*>#&saѢ[z73wNN1s(ǛƗGw׀] q򏞆yTTv!@q޴cɪRj4]=~sS%\J H^`s\ [J <{v# r @"Χ3kDQGFܥZ I4r\t ]Niԓ{Q-ùD!,7F xeKu8Ni*;<;|P*c׹!'տ"7(lm$aΏۻx#G~”kk(e)PKQ9"L.5)zJ۹06`tsa݊rr&K!3he^_ BgEU{0"7K+L6nAN( Tāb;-'yGj HyW>OgY3$|: klX.eUg_W:V(e3i&-,*J$S9"GtsY2Ze)]nR@WW{PhM&R. O@UhZ\[kܣt #Xʕ\;je>sXr :Fl;!%;X#(}X@\%;An38rWMoOɮA vB;iĺGnI0h4N.п{$4%gv4) A3ɧN"cj9[}[~+?5m5V ,e׍:'r;'[T(:3L}\^|Sj;ԙ[ rP+e\N/r2( ̝oItslڿK~~R2CeDRzp ֕rYހHzfaFW5SȝȖ +q;:oş,b!eV^Ƒmɤ|^K#&`D1Q )dʁNC9E84f&R1P'1mD3xmyѭspb`+[͟R1P'1휻eZCGt}Aft!9eHK}wzc=ƸK*KUk uvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003520751515134272254017712 0ustar rootrootJan 22 00:07:03 crc systemd[1]: Starting Kubernetes Kubelet... Jan 22 00:07:03 crc restorecon[4702]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:03 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 00:07:04 crc restorecon[4702]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 22 00:07:04 crc kubenswrapper[4800]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.646009 4800 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649683 4800 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649707 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649713 4800 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649720 4800 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649727 4800 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649733 4800 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649740 4800 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649747 4800 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649753 4800 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649759 4800 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649765 4800 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649771 4800 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649776 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649782 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649787 4800 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649795 4800 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649801 4800 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649808 4800 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649828 4800 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649834 4800 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649841 4800 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649846 4800 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649851 4800 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649857 4800 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649862 4800 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649868 4800 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649875 4800 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649880 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649906 4800 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649911 4800 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649917 4800 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649922 4800 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649927 4800 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649932 4800 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649937 4800 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649943 4800 feature_gate.go:330] unrecognized feature gate: Example Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649948 4800 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649953 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649958 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649964 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649970 4800 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649975 4800 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649980 4800 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649986 4800 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649991 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.649996 4800 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650001 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650006 4800 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650012 4800 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650017 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650022 4800 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650028 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650034 4800 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650039 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650045 4800 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650050 4800 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650059 4800 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650066 4800 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650073 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650078 4800 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650084 4800 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650090 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650095 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650100 4800 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650106 4800 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650112 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650119 4800 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650126 4800 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650132 4800 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650138 4800 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.650143 4800 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650586 4800 flags.go:64] FLAG: --address="0.0.0.0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650601 4800 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650611 4800 flags.go:64] FLAG: --anonymous-auth="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650619 4800 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650627 4800 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650633 4800 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650641 4800 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650649 4800 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650655 4800 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650661 4800 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650667 4800 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650674 4800 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650680 4800 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650687 4800 flags.go:64] FLAG: --cgroup-root="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650693 4800 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650700 4800 flags.go:64] FLAG: --client-ca-file="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650707 4800 flags.go:64] FLAG: --cloud-config="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650713 4800 flags.go:64] FLAG: --cloud-provider="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650719 4800 flags.go:64] FLAG: --cluster-dns="[]" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650728 4800 flags.go:64] FLAG: --cluster-domain="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650734 4800 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650740 4800 flags.go:64] FLAG: --config-dir="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650747 4800 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650753 4800 flags.go:64] FLAG: --container-log-max-files="5" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650761 4800 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650767 4800 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650773 4800 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650779 4800 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650785 4800 flags.go:64] FLAG: --contention-profiling="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650791 4800 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650797 4800 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650803 4800 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650811 4800 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650819 4800 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650825 4800 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650831 4800 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650837 4800 flags.go:64] FLAG: --enable-load-reader="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650843 4800 flags.go:64] FLAG: --enable-server="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650850 4800 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650858 4800 flags.go:64] FLAG: --event-burst="100" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650864 4800 flags.go:64] FLAG: --event-qps="50" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650871 4800 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650877 4800 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650901 4800 flags.go:64] FLAG: --eviction-hard="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650910 4800 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650916 4800 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650922 4800 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650928 4800 flags.go:64] FLAG: --eviction-soft="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650935 4800 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650941 4800 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650947 4800 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650953 4800 flags.go:64] FLAG: --experimental-mounter-path="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650959 4800 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650966 4800 flags.go:64] FLAG: --fail-swap-on="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650971 4800 flags.go:64] FLAG: --feature-gates="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650979 4800 flags.go:64] FLAG: --file-check-frequency="20s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650985 4800 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650991 4800 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.650997 4800 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651003 4800 flags.go:64] FLAG: --healthz-port="10248" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651012 4800 flags.go:64] FLAG: --help="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651018 4800 flags.go:64] FLAG: --hostname-override="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651024 4800 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651030 4800 flags.go:64] FLAG: --http-check-frequency="20s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651038 4800 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651045 4800 flags.go:64] FLAG: --image-credential-provider-config="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651051 4800 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651057 4800 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651064 4800 flags.go:64] FLAG: --image-service-endpoint="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651071 4800 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651077 4800 flags.go:64] FLAG: --kube-api-burst="100" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651083 4800 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651089 4800 flags.go:64] FLAG: --kube-api-qps="50" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651095 4800 flags.go:64] FLAG: --kube-reserved="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651102 4800 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651107 4800 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651113 4800 flags.go:64] FLAG: --kubelet-cgroups="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651120 4800 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651126 4800 flags.go:64] FLAG: --lock-file="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651132 4800 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651138 4800 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651144 4800 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651152 4800 flags.go:64] FLAG: --log-json-split-stream="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651158 4800 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651165 4800 flags.go:64] FLAG: --log-text-split-stream="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651171 4800 flags.go:64] FLAG: --logging-format="text" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651177 4800 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651184 4800 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651190 4800 flags.go:64] FLAG: --manifest-url="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651196 4800 flags.go:64] FLAG: --manifest-url-header="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651204 4800 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651210 4800 flags.go:64] FLAG: --max-open-files="1000000" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651218 4800 flags.go:64] FLAG: --max-pods="110" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651225 4800 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651231 4800 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651238 4800 flags.go:64] FLAG: --memory-manager-policy="None" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651244 4800 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651250 4800 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651256 4800 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651262 4800 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651276 4800 flags.go:64] FLAG: --node-status-max-images="50" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651282 4800 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651288 4800 flags.go:64] FLAG: --oom-score-adj="-999" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651295 4800 flags.go:64] FLAG: --pod-cidr="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651302 4800 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651310 4800 flags.go:64] FLAG: --pod-manifest-path="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651316 4800 flags.go:64] FLAG: --pod-max-pids="-1" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651322 4800 flags.go:64] FLAG: --pods-per-core="0" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651329 4800 flags.go:64] FLAG: --port="10250" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651335 4800 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651341 4800 flags.go:64] FLAG: --provider-id="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651347 4800 flags.go:64] FLAG: --qos-reserved="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651353 4800 flags.go:64] FLAG: --read-only-port="10255" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651359 4800 flags.go:64] FLAG: --register-node="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651365 4800 flags.go:64] FLAG: --register-schedulable="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651371 4800 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651381 4800 flags.go:64] FLAG: --registry-burst="10" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651413 4800 flags.go:64] FLAG: --registry-qps="5" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651420 4800 flags.go:64] FLAG: --reserved-cpus="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651426 4800 flags.go:64] FLAG: --reserved-memory="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651435 4800 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651442 4800 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651449 4800 flags.go:64] FLAG: --rotate-certificates="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651456 4800 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651463 4800 flags.go:64] FLAG: --runonce="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651470 4800 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651477 4800 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651484 4800 flags.go:64] FLAG: --seccomp-default="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651490 4800 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651498 4800 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651506 4800 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651513 4800 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651520 4800 flags.go:64] FLAG: --storage-driver-password="root" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651527 4800 flags.go:64] FLAG: --storage-driver-secure="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651534 4800 flags.go:64] FLAG: --storage-driver-table="stats" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651541 4800 flags.go:64] FLAG: --storage-driver-user="root" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651548 4800 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651556 4800 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651563 4800 flags.go:64] FLAG: --system-cgroups="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651570 4800 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651581 4800 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651588 4800 flags.go:64] FLAG: --tls-cert-file="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651594 4800 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651603 4800 flags.go:64] FLAG: --tls-min-version="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651609 4800 flags.go:64] FLAG: --tls-private-key-file="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651615 4800 flags.go:64] FLAG: --topology-manager-policy="none" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651623 4800 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651629 4800 flags.go:64] FLAG: --topology-manager-scope="container" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651636 4800 flags.go:64] FLAG: --v="2" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651644 4800 flags.go:64] FLAG: --version="false" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651652 4800 flags.go:64] FLAG: --vmodule="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651659 4800 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.651666 4800 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651812 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651818 4800 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651824 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651832 4800 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651839 4800 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651844 4800 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651850 4800 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651855 4800 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651861 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651867 4800 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651872 4800 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651878 4800 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651883 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651909 4800 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651915 4800 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651920 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651925 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651930 4800 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651936 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651941 4800 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651946 4800 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651951 4800 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651956 4800 feature_gate.go:330] unrecognized feature gate: Example Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651962 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651968 4800 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651974 4800 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651979 4800 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651984 4800 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651989 4800 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.651995 4800 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652000 4800 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652005 4800 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652010 4800 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652015 4800 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652022 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652027 4800 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652033 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652038 4800 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652043 4800 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652048 4800 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652055 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652061 4800 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652066 4800 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652071 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652076 4800 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652082 4800 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652087 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652092 4800 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652097 4800 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652102 4800 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652108 4800 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652113 4800 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652120 4800 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652127 4800 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652134 4800 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652140 4800 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652145 4800 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652151 4800 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652157 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652164 4800 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652169 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652174 4800 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652179 4800 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652185 4800 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652190 4800 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652196 4800 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652201 4800 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652206 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652213 4800 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652220 4800 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.652227 4800 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.652247 4800 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.662393 4800 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.662444 4800 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662530 4800 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662542 4800 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662548 4800 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662554 4800 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662558 4800 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662563 4800 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662567 4800 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662572 4800 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662579 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662585 4800 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662590 4800 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662596 4800 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662601 4800 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662605 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662611 4800 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662616 4800 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662620 4800 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662625 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662631 4800 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662640 4800 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662646 4800 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662651 4800 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662656 4800 feature_gate.go:330] unrecognized feature gate: Example Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662661 4800 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662666 4800 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662670 4800 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662675 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662679 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662683 4800 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662687 4800 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662692 4800 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662696 4800 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662701 4800 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662705 4800 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662710 4800 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662717 4800 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662722 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662727 4800 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662733 4800 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662738 4800 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662743 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662747 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662752 4800 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662756 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662762 4800 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662766 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662770 4800 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662774 4800 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662780 4800 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662787 4800 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662794 4800 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662800 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662807 4800 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662813 4800 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662818 4800 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662823 4800 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662838 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662842 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662847 4800 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662851 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662855 4800 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662860 4800 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662864 4800 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662869 4800 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662873 4800 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662878 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662882 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662904 4800 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662909 4800 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662913 4800 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.662918 4800 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.662927 4800 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663069 4800 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663077 4800 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663082 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663088 4800 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663093 4800 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663096 4800 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663102 4800 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663105 4800 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663109 4800 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663113 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663117 4800 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663120 4800 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663124 4800 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663128 4800 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663133 4800 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663137 4800 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663141 4800 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663145 4800 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663149 4800 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663152 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663156 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663159 4800 feature_gate.go:330] unrecognized feature gate: Example Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663163 4800 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663166 4800 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663170 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663173 4800 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663177 4800 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663181 4800 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663185 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663190 4800 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663194 4800 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663200 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663204 4800 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663208 4800 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663212 4800 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663217 4800 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663221 4800 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663226 4800 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663232 4800 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663237 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663242 4800 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663246 4800 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663251 4800 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663255 4800 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663261 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663266 4800 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663271 4800 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663275 4800 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663281 4800 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663286 4800 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663291 4800 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663296 4800 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663300 4800 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663305 4800 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663310 4800 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663315 4800 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663319 4800 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663324 4800 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663328 4800 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663333 4800 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663507 4800 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663516 4800 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663521 4800 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663526 4800 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663529 4800 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663534 4800 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663538 4800 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663542 4800 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663547 4800 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663552 4800 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.663556 4800 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.663562 4800 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.663742 4800 server.go:940] "Client rotation is on, will bootstrap in background" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.666065 4800 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.666154 4800 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.666605 4800 server.go:997] "Starting client certificate rotation" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.666638 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.667013 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-11 12:55:28.325119639 +0000 UTC Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.667096 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.671520 4800 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.675648 4800 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.677109 4800 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.688588 4800 log.go:25] "Validated CRI v1 runtime API" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.703806 4800 log.go:25] "Validated CRI v1 image API" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.705789 4800 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.708777 4800 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-22-00-01-28-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.708830 4800 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.739614 4800 manager.go:217] Machine: {Timestamp:2026-01-22 00:07:04.737493749 +0000 UTC m=+0.217796867 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f BootID:929f49ac-7e30-48d3-8f26-0c30c68b0fdc Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:65:50:a6 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:65:50:a6 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:2b:51:5e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:11:3a:c7 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:15:43:f9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a8:1d:db Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:c2:a3:92:56:ae Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e6:dd:25:89:7e:96 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.740109 4800 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.740472 4800 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.741386 4800 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.741672 4800 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.741722 4800 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.743939 4800 topology_manager.go:138] "Creating topology manager with none policy" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.743992 4800 container_manager_linux.go:303] "Creating device plugin manager" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.744500 4800 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.744714 4800 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.745203 4800 state_mem.go:36] "Initialized new in-memory state store" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.745934 4800 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.746837 4800 kubelet.go:418] "Attempting to sync node with API server" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.746910 4800 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.746945 4800 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.746969 4800 kubelet.go:324] "Adding apiserver pod source" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.746996 4800 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.749438 4800 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.749763 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.749864 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.750063 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.750137 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.750263 4800 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.750958 4800 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751549 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751585 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751596 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751607 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751622 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751632 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751641 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751657 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751668 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751677 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751691 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751701 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.751874 4800 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.752531 4800 server.go:1280] "Started kubelet" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.752846 4800 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.752958 4800 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.753138 4800 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.753779 4800 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.755039 4800 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.242:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188ce4e4bead7206 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 00:07:04.752476678 +0000 UTC m=+0.232779726,LastTimestamp:2026-01-22 00:07:04.752476678 +0000 UTC m=+0.232779726,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 00:07:04 crc systemd[1]: Started Kubernetes Kubelet. Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.756007 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.756049 4800 server.go:460] "Adding debug handlers to kubelet server" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.756084 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 14:25:52.458398283 +0000 UTC Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.756053 4800 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.757208 4800 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.757230 4800 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.757203 4800 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.757311 4800 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.760987 4800 factory.go:55] Registering systemd factory Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.764917 4800 factory.go:221] Registration of the systemd container factory successfully Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.761068 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.765818 4800 factory.go:153] Registering CRI-O factory Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.765940 4800 factory.go:221] Registration of the crio container factory successfully Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.769003 4800 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.769129 4800 factory.go:103] Registering Raw factory Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.769220 4800 manager.go:1196] Started watching for new ooms in manager Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.768998 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.764331 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="200ms" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.770309 4800 manager.go:319] Starting recovery of all containers Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772371 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772442 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772488 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772503 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772525 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772538 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772551 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772564 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772583 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772600 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772640 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772656 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772669 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772691 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772705 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772719 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772754 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772769 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772782 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772796 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772809 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772825 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772842 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772856 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772876 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772909 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772928 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772943 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772957 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772974 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.772990 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773006 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773020 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773037 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773050 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773063 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773078 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773093 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773107 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773122 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773161 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773174 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773191 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773206 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773222 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773236 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773250 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773266 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773281 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773296 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773310 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773326 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773345 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773388 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773407 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773424 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773439 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773460 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773474 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773487 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773501 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773515 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773530 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773545 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773560 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773574 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773589 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773644 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773659 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773673 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773688 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773704 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773717 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773731 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773745 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773757 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773779 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773795 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773809 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773823 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773836 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773851 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773865 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773879 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773912 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773926 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773939 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773953 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773966 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.773985 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774000 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774016 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774030 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774044 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774058 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774072 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774087 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774100 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774115 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774128 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774142 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774156 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774170 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774184 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774206 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774223 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774237 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774254 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774267 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774283 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774299 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774315 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774329 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774343 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774357 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774370 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774382 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774396 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774408 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774422 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774436 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774455 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774468 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774483 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774497 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774514 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774529 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774545 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774558 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774571 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774585 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774600 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774615 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774631 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774646 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774660 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774674 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774687 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774700 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774716 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774730 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774744 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774757 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774773 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774787 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774800 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774814 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774827 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774840 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774852 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774866 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774880 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774912 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774928 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774942 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774958 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774974 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.774988 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775004 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775017 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775032 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775047 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775061 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775075 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775091 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775106 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.775122 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.778936 4800 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.778972 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.778997 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.779013 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.779028 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782193 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782209 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782230 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782247 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782260 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782273 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782287 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782301 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782314 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782330 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782343 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782357 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782370 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782385 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782398 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782414 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782426 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782440 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782454 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782471 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782487 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782503 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782528 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782542 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782556 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782571 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782584 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782599 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782613 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782627 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782639 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782656 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782670 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782684 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782697 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782711 4800 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782724 4800 reconstruct.go:97] "Volume reconstruction finished" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.782734 4800 reconciler.go:26] "Reconciler: start to sync state" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.786171 4800 manager.go:324] Recovery completed Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.796393 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.800190 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.800436 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.800537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.805869 4800 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.805929 4800 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.805962 4800 state_mem.go:36] "Initialized new in-memory state store" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.813433 4800 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.816591 4800 policy_none.go:49] "None policy: Start" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.816748 4800 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.816795 4800 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.816837 4800 kubelet.go:2335] "Starting kubelet main sync loop" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.816918 4800 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 22 00:07:04 crc kubenswrapper[4800]: W0122 00:07:04.818440 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.818602 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.819074 4800 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.819112 4800 state_mem.go:35] "Initializing new in-memory state store" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.857853 4800 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.876589 4800 manager.go:334] "Starting Device Plugin manager" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.877721 4800 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.877740 4800 server.go:79] "Starting device plugin registration server" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.878336 4800 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.878358 4800 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.878941 4800 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.879021 4800 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.879034 4800 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.892274 4800 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.918154 4800 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.918258 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919379 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919390 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919516 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919837 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.919938 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921047 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921079 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921163 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921193 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921303 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921547 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921575 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921941 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921973 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.921984 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922091 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922274 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922301 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922311 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922500 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922576 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922941 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922983 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.922997 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.923208 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.923478 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.923516 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924510 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924523 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924636 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924653 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924616 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.924913 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.925106 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.925809 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.926029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.926128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.970845 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="400ms" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.978984 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.980502 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.980561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.980574 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.980639 4800 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 00:07:04 crc kubenswrapper[4800]: E0122 00:07:04.981417 4800 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.242:6443: connect: connection refused" node="crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985274 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985308 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985334 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985350 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985366 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985379 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985407 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985423 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985441 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985456 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985471 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985485 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985500 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985518 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:04 crc kubenswrapper[4800]: I0122 00:07:04.985556 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086616 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086722 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086768 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086812 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086846 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086878 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086934 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087020 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087078 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087039 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086953 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087144 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087146 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087171 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.086957 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087203 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087227 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087235 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087271 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087271 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087310 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087328 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087356 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087366 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087381 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087414 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087445 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087510 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087553 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.087414 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.182475 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.184541 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.184665 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.184741 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.184834 4800 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 00:07:05 crc kubenswrapper[4800]: E0122 00:07:05.185401 4800 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.242:6443: connect: connection refused" node="crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.269293 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.283272 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: W0122 00:07:05.297682 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-930fbd252f65e347639213c8443a804afaacce241f8d0d8719162848ccbcf4f3 WatchSource:0}: Error finding container 930fbd252f65e347639213c8443a804afaacce241f8d0d8719162848ccbcf4f3: Status 404 returned error can't find the container with id 930fbd252f65e347639213c8443a804afaacce241f8d0d8719162848ccbcf4f3 Jan 22 00:07:05 crc kubenswrapper[4800]: W0122 00:07:05.303233 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-475508737c93e1846abe471c9b19c0aa042ecc33d698d467c36428e886a6e7d9 WatchSource:0}: Error finding container 475508737c93e1846abe471c9b19c0aa042ecc33d698d467c36428e886a6e7d9: Status 404 returned error can't find the container with id 475508737c93e1846abe471c9b19c0aa042ecc33d698d467c36428e886a6e7d9 Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.318863 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.333252 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.341179 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:05 crc kubenswrapper[4800]: W0122 00:07:05.360537 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-4e7c8cac88b409ad44fa9da398dab8f29e4f12f9237fabd6ef1a6fa8326d56d6 WatchSource:0}: Error finding container 4e7c8cac88b409ad44fa9da398dab8f29e4f12f9237fabd6ef1a6fa8326d56d6: Status 404 returned error can't find the container with id 4e7c8cac88b409ad44fa9da398dab8f29e4f12f9237fabd6ef1a6fa8326d56d6 Jan 22 00:07:05 crc kubenswrapper[4800]: W0122 00:07:05.366295 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-15bd2f35a456167820233b2a2a753a5d244dd23926f940bf1e469f37f03f0c35 WatchSource:0}: Error finding container 15bd2f35a456167820233b2a2a753a5d244dd23926f940bf1e469f37f03f0c35: Status 404 returned error can't find the container with id 15bd2f35a456167820233b2a2a753a5d244dd23926f940bf1e469f37f03f0c35 Jan 22 00:07:05 crc kubenswrapper[4800]: E0122 00:07:05.371720 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="800ms" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.586423 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.588603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.588670 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.588685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.588727 4800 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 00:07:05 crc kubenswrapper[4800]: E0122 00:07:05.589384 4800 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.242:6443: connect: connection refused" node="crc" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.754184 4800 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.756386 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 11:37:54.545376107 +0000 UTC Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.824204 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27" exitCode=0 Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.824277 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.824366 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4e7c8cac88b409ad44fa9da398dab8f29e4f12f9237fabd6ef1a6fa8326d56d6"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.824455 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.825791 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.825824 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.825833 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.827511 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.827540 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32aca283c4fa92ca87a07b6cfc78d4aed37d3a374d50280bb7e1cf7cbb8377a5"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.828197 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.829031 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.829057 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.829066 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.829983 4800 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893" exitCode=0 Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.830038 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.830053 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"475508737c93e1846abe471c9b19c0aa042ecc33d698d467c36428e886a6e7d9"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.830134 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.831356 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.831379 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.831387 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.832214 4800 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6fbf8ccaf5f4e8dce2e62baaf790a859bdfa4a86f2eb510a85ba0f824492273e" exitCode=0 Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.832269 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6fbf8ccaf5f4e8dce2e62baaf790a859bdfa4a86f2eb510a85ba0f824492273e"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.832288 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"930fbd252f65e347639213c8443a804afaacce241f8d0d8719162848ccbcf4f3"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.832334 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.833019 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.833044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.833054 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.834276 4800 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837" exitCode=0 Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.834298 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.834313 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"15bd2f35a456167820233b2a2a753a5d244dd23926f940bf1e469f37f03f0c35"} Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.834378 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.835657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.835845 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:05 crc kubenswrapper[4800]: I0122 00:07:05.835866 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: W0122 00:07:06.054757 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:06 crc kubenswrapper[4800]: E0122 00:07:06.054869 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:06 crc kubenswrapper[4800]: W0122 00:07:06.164897 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:06 crc kubenswrapper[4800]: E0122 00:07:06.164989 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:06 crc kubenswrapper[4800]: E0122 00:07:06.174819 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="1.6s" Jan 22 00:07:06 crc kubenswrapper[4800]: W0122 00:07:06.198582 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:06 crc kubenswrapper[4800]: E0122 00:07:06.198727 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:06 crc kubenswrapper[4800]: W0122 00:07:06.318704 4800 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:07:06 crc kubenswrapper[4800]: E0122 00:07:06.318791 4800 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.242:6443: connect: connection refused" logger="UnhandledError" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.389812 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.391505 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.391552 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.391561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.391599 4800 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.747945 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.757195 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 13:23:46.933224044 +0000 UTC Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.838754 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.838815 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.838825 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.838834 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.841206 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.841231 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.841240 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.841336 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.842481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.842513 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.842526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.843113 4800 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b" exitCode=0 Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.843187 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.843339 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.844176 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.844201 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.844214 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.845637 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5b954f5a0981a1434fee0dc9162ca3d565f303c71c42a4e5c57fdd8f6823aec5"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.845700 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.846730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.846772 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.846790 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.848431 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.848455 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.848472 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883"} Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.848524 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.861797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.862070 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.862083 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.923906 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.979905 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:06 crc kubenswrapper[4800]: I0122 00:07:06.986552 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.758329 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 20:12:01.46913783 +0000 UTC Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.858513 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003"} Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.858686 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.860481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.860514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.860529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.863054 4800 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c" exitCode=0 Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.863149 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c"} Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.863285 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.863339 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.863291 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.865459 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.865509 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.865528 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866072 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866114 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866132 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866744 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866798 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.866821 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:07 crc kubenswrapper[4800]: I0122 00:07:07.932473 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.398943 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.759640 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 01:21:12.321538442 +0000 UTC Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872063 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8"} Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872134 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872168 4800 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872230 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872142 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4"} Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872271 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0"} Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872355 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.872376 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1"} Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.873372 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.873453 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.873475 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.874107 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.874149 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:08 crc kubenswrapper[4800]: I0122 00:07:08.874164 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.101171 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.760870 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 02:06:07.175152282 +0000 UTC Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.833474 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.881462 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3"} Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.881583 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.881583 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.882752 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887021 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887094 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887120 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887468 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887498 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887567 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.887853 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.888112 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:09 crc kubenswrapper[4800]: I0122 00:07:09.888293 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.762038 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 11:41:02.903068641 +0000 UTC Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.884937 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.884937 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.886862 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.886979 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.887008 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.887010 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.887165 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:10 crc kubenswrapper[4800]: I0122 00:07:10.887191 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:11 crc kubenswrapper[4800]: I0122 00:07:11.763186 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 22:56:00.02598068 +0000 UTC Jan 22 00:07:12 crc kubenswrapper[4800]: I0122 00:07:12.763368 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 14:57:48.397940945 +0000 UTC Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.472830 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.473613 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.475655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.475733 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.475745 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:13 crc kubenswrapper[4800]: I0122 00:07:13.764013 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 16:19:36.784889584 +0000 UTC Jan 22 00:07:14 crc kubenswrapper[4800]: I0122 00:07:14.764376 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 10:58:01.285093971 +0000 UTC Jan 22 00:07:14 crc kubenswrapper[4800]: E0122 00:07:14.893405 4800 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 22 00:07:15 crc kubenswrapper[4800]: I0122 00:07:15.764768 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 10:22:04.955371702 +0000 UTC Jan 22 00:07:16 crc kubenswrapper[4800]: E0122 00:07:16.392549 4800 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.550929 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.551307 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.553416 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.553488 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.553518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.555623 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:16 crc kubenswrapper[4800]: E0122 00:07:16.749910 4800 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.755801 4800 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.765415 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 19:37:46.994206646 +0000 UTC Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.901950 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.902797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.902835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:16 crc kubenswrapper[4800]: I0122 00:07:16.902844 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.104482 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.104710 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.106133 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.106187 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.106200 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.587643 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.587748 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.594478 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.594567 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.766113 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 16:49:05.897562571 +0000 UTC Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.942556 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]log ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]etcd ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/generic-apiserver-start-informers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-filter ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-apiextensions-informers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-apiextensions-controllers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/crd-informer-synced ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-system-namespaces-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 22 00:07:17 crc kubenswrapper[4800]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/bootstrap-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/start-kube-aggregator-informers ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-registration-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-discovery-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]autoregister-completion ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-openapi-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 22 00:07:17 crc kubenswrapper[4800]: livez check failed Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.942672 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.992969 4800 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.995318 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.995363 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.995376 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:17 crc kubenswrapper[4800]: I0122 00:07:17.995405 4800 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 00:07:18 crc kubenswrapper[4800]: I0122 00:07:18.766769 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 10:11:33.612597272 +0000 UTC Jan 22 00:07:19 crc kubenswrapper[4800]: I0122 00:07:19.550759 4800 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 00:07:19 crc kubenswrapper[4800]: I0122 00:07:19.550862 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 00:07:19 crc kubenswrapper[4800]: I0122 00:07:19.767467 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 14:16:37.029141152 +0000 UTC Jan 22 00:07:20 crc kubenswrapper[4800]: I0122 00:07:20.767830 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 22:17:51.163383484 +0000 UTC Jan 22 00:07:20 crc kubenswrapper[4800]: I0122 00:07:20.773018 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 00:07:20 crc kubenswrapper[4800]: I0122 00:07:20.791495 4800 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 00:07:20 crc kubenswrapper[4800]: I0122 00:07:20.817832 4800 csr.go:261] certificate signing request csr-9kb6l is approved, waiting to be issued Jan 22 00:07:20 crc kubenswrapper[4800]: I0122 00:07:20.825234 4800 csr.go:257] certificate signing request csr-9kb6l is issued Jan 22 00:07:21 crc kubenswrapper[4800]: I0122 00:07:21.768415 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 18:12:53.623786518 +0000 UTC Jan 22 00:07:21 crc kubenswrapper[4800]: I0122 00:07:21.826809 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-22 00:02:20 +0000 UTC, rotation deadline is 2026-12-09 17:58:39.936132247 +0000 UTC Jan 22 00:07:21 crc kubenswrapper[4800]: I0122 00:07:21.826857 4800 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7721h51m18.109280438s for next certificate rotation Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.578978 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580458 4800 trace.go:236] Trace[708789568]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 00:07:08.272) (total time: 14307ms): Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[708789568]: ---"Objects listed" error: 14307ms (00:07:22.580) Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[708789568]: [14.307418914s] [14.307418914s] END Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580503 4800 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580874 4800 trace.go:236] Trace[1967022521]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 00:07:07.932) (total time: 14648ms): Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[1967022521]: ---"Objects listed" error: 14648ms (00:07:22.580) Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[1967022521]: [14.648474399s] [14.648474399s] END Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580916 4800 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580922 4800 trace.go:236] Trace[1591549360]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 00:07:09.383) (total time: 13197ms): Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[1591549360]: ---"Objects listed" error: 13197ms (00:07:22.580) Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[1591549360]: [13.197340547s] [13.197340547s] END Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.580943 4800 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.587985 4800 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.590014 4800 trace.go:236] Trace[564849073]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 00:07:09.003) (total time: 13586ms): Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[564849073]: ---"Objects listed" error: 13584ms (00:07:22.588) Jan 22 00:07:22 crc kubenswrapper[4800]: Trace[564849073]: [13.586441066s] [13.586441066s] END Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.590045 4800 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.630785 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57740->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.630871 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57740->192.168.126.11:17697: read: connection reset by peer" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.759854 4800 apiserver.go:52] "Watching apiserver" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.762563 4800 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.762967 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.763479 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.763595 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.763744 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.763988 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.764044 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.764231 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.764822 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.764979 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.765073 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.766762 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.766988 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.767217 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768046 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768076 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768047 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768188 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768349 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.768530 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 04:07:29.868626043 +0000 UTC Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.769332 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.807140 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.823333 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.841334 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.858160 4800 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.861273 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-fp7s5"] Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.861766 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.863596 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.863978 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.864719 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.866640 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.884057 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889571 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889626 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889655 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889679 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889707 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889727 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889747 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889778 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889804 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889826 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889849 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889871 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889909 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889935 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889959 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.889981 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890007 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890034 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890063 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890090 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890116 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890139 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890164 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890190 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890217 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890242 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890266 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890326 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890351 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890376 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890398 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890424 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890451 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890475 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890505 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890527 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890550 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890575 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890597 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890621 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890652 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890676 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890702 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890736 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890766 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890793 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890820 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890854 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890878 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890923 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.890949 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891108 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891135 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891160 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891184 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891207 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891236 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891260 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891286 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891314 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891338 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891361 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891408 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891433 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891458 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891486 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891511 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891536 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891567 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891591 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891617 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891642 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891668 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891694 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891719 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891747 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891772 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891796 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891821 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891849 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891875 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891916 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891942 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.891975 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892004 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892029 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892051 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892075 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892103 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892126 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892148 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892173 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892197 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892222 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892248 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892273 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892297 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892323 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892348 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892373 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892399 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892424 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892450 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892476 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892503 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892528 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892551 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892574 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892598 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892626 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892651 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892681 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892706 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892732 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892757 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892780 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892809 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892836 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892863 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892905 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892943 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892969 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.892996 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893021 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893048 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893073 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893099 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893125 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893150 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893175 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893201 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893225 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893251 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893278 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893305 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893329 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893363 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893386 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893410 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893436 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893462 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893489 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893529 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893556 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893582 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893611 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893638 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893662 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893687 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893712 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893736 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893763 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893791 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893820 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893846 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893872 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893916 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893947 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.893977 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894008 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894038 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894068 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894093 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894118 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894145 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894172 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894200 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894226 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894254 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894279 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894306 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894335 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894362 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894387 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894414 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894439 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894463 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894489 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894513 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894538 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894565 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894591 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894618 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894645 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894677 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894705 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894735 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894761 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894788 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894816 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894843 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894868 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894909 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894936 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.894961 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895029 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895062 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895096 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895126 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895156 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895185 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895218 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895247 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895273 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895298 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895326 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895354 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895382 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.895431 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896103 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896446 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896472 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896536 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896652 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896704 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896930 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.896956 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897141 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897300 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897545 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897745 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897847 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.897942 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898024 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898187 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898507 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898531 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898725 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898760 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898778 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898923 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.898955 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.899072 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.899101 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901110 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901121 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901203 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901227 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901243 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901335 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901414 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901453 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901533 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901544 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901607 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901679 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901708 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901723 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.901964 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902108 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902152 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902233 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902304 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902412 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902446 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902635 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902744 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902958 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.902978 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903115 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903453 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903549 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903641 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903808 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903921 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.903947 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904186 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904233 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904358 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904379 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904436 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904545 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904714 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.904933 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905017 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905235 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905281 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905286 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905361 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905406 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905931 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.905984 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906153 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906221 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906288 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906415 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906558 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906622 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906742 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.906945 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907094 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907277 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907473 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907500 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907731 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.907786 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908233 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908229 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908307 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908416 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908503 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.908596 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.908631 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.908639 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.908650 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.908735 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:23.408707828 +0000 UTC m=+18.889010876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.909212 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.909454 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.909786 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.909870 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910003 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910170 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910238 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910260 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910465 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.910797 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911342 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911395 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911770 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911833 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911931 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911937 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.911962 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912080 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912313 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912472 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912523 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912635 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912811 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.912979 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.913281 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.914364 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.914651 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915112 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915243 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915304 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915368 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915227 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915751 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915817 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.915875 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.916160 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.916489 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.917981 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.918041 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.918268 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.918389 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.918443 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.918533 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:23.418500952 +0000 UTC m=+18.898804000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.919221 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.919696 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.920314 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.920336 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.921974 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.922126 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.922940 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.923411 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.923594 4800 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.923904 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.924232 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.924671 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.924929 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.925060 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.926109 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.926442 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.926543 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.927321 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.927442 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.927770 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928072 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928340 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928387 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928494 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928563 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.928782 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.929954 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930048 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930131 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930325 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930343 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930819 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930423 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930447 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930746 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.930756 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.931225 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.931341 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:23.43131195 +0000 UTC m=+18.911615198 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.931637 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.931748 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.931766 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.931755 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.931806 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:23.431796814 +0000 UTC m=+18.912099852 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.932097 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.932238 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.932420 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.932846 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.933125 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.933294 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.933349 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.935833 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.936294 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.936451 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.937472 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.938525 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.939155 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.939830 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.940198 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.940264 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.944565 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.946422 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.947310 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.948327 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003" exitCode=255 Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.948438 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003"} Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.949416 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.957278 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.957925 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.958023 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.958098 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:22 crc kubenswrapper[4800]: E0122 00:07:22.958216 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:23.458185512 +0000 UTC m=+18.938488550 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.960482 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.961530 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.962162 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.962239 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.964569 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.969267 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.971515 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.976209 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.989210 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.989499 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.989762 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.992072 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.992772 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.996838 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b37dab4b-08e1-4daa-86d9-561432a82428-hosts-file\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.996898 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.996934 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45ldt\" (UniqueName: \"kubernetes.io/projected/b37dab4b-08e1-4daa-86d9-561432a82428-kube-api-access-45ldt\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.996990 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997063 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997075 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997085 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997098 4800 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997109 4800 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997118 4800 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997129 4800 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997142 4800 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997152 4800 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997162 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997171 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997182 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997193 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997202 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997211 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997222 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997239 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997248 4800 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997260 4800 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:22 crc kubenswrapper[4800]: I0122 00:07:22.997269 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997277 4800 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997287 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997298 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997310 4800 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997319 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997330 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997342 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997351 4800 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997361 4800 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997370 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997381 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997390 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997399 4800 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997410 4800 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997420 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997429 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997438 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997450 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997460 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997469 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997480 4800 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997499 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997507 4800 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997517 4800 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997529 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997538 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997546 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997555 4800 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997570 4800 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997578 4800 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997593 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997607 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997624 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997638 4800 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997650 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997637 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997661 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997746 4800 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997762 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997771 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997777 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997800 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997813 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997823 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997855 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997869 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997900 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997911 4800 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997920 4800 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997930 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997941 4800 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997951 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.997960 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998035 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998060 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998074 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998099 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998116 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998127 4800 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998138 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998148 4800 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998160 4800 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998171 4800 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998180 4800 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998188 4800 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998201 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998211 4800 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998221 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998236 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998247 4800 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998262 4800 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998271 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998283 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998291 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998328 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998340 4800 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998353 4800 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998363 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998373 4800 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998385 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998401 4800 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998413 4800 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998422 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998434 4800 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998443 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998452 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998463 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998474 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998483 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998492 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998501 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998513 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998522 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998531 4800 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998542 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998551 4800 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998560 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998570 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998582 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998593 4800 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998602 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998612 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998627 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998637 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998645 4800 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998657 4800 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998666 4800 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998675 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998684 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998695 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998706 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998715 4800 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998726 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998737 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998747 4800 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998756 4800 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998765 4800 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998777 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998785 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998794 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998805 4800 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998813 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998822 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998832 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998844 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998853 4800 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998862 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998871 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998899 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998908 4800 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998917 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998927 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998939 4800 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998948 4800 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998958 4800 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998970 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998979 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998989 4800 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.998998 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999010 4800 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999021 4800 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999030 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999040 4800 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999052 4800 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999061 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999070 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999082 4800 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999090 4800 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999099 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999108 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999118 4800 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999127 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999138 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999147 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999158 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999167 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999177 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999187 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999198 4800 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999206 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999215 4800 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999227 4800 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999237 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999247 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999256 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999267 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999278 4800 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999287 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999297 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999309 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999317 4800 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:22.999326 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.007814 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.008983 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.011734 4800 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.011997 4800 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.013240 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.013274 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.013284 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.013302 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.013313 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.018438 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.025666 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.028493 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.031161 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.038561 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.046793 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.047104 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.047194 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.047300 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.047377 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.048881 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.065969 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.068108 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.069561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.069649 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.069736 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.069813 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.069876 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.078596 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.084701 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.086469 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.095757 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.099837 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b37dab4b-08e1-4daa-86d9-561432a82428-hosts-file\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.099917 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ldt\" (UniqueName: \"kubernetes.io/projected/b37dab4b-08e1-4daa-86d9-561432a82428-kube-api-access-45ldt\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.099983 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.100009 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.100022 4800 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.100039 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.100059 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b37dab4b-08e1-4daa-86d9-561432a82428-hosts-file\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:23 crc kubenswrapper[4800]: W0122 00:07:23.105356 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-10a08d74f9cace0b37f170da8fcfb383584ffc3a20fcc89dae715698a97c9541 WatchSource:0}: Error finding container 10a08d74f9cace0b37f170da8fcfb383584ffc3a20fcc89dae715698a97c9541: Status 404 returned error can't find the container with id 10a08d74f9cace0b37f170da8fcfb383584ffc3a20fcc89dae715698a97c9541 Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109162 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109221 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109281 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.109688 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: W0122 00:07:23.120626 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-ab311420ac3b38ff4bac59dce09846554d3c55521e90168b65c7ed74e3ecb471 WatchSource:0}: Error finding container ab311420ac3b38ff4bac59dce09846554d3c55521e90168b65c7ed74e3ecb471: Status 404 returned error can't find the container with id ab311420ac3b38ff4bac59dce09846554d3c55521e90168b65c7ed74e3ecb471 Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.132650 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45ldt\" (UniqueName: \"kubernetes.io/projected/b37dab4b-08e1-4daa-86d9-561432a82428-kube-api-access-45ldt\") pod \"node-resolver-fp7s5\" (UID: \"b37dab4b-08e1-4daa-86d9-561432a82428\") " pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.132771 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.140602 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.145910 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.145957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.145969 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.145991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.146006 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.163407 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.163527 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173174 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173243 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173271 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173286 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.173475 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fp7s5" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.276961 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-bv9wt"] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.278019 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vfgd4"] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.278220 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.278238 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.281109 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.281137 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.281145 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.281158 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.281168 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.288883 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.288988 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-mrfxg"] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289137 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289328 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289358 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289495 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289559 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289667 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.289812 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.294344 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.294692 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.295254 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.295524 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.295784 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.300067 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.321273 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.334372 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.346914 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.363139 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.384330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.384396 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.384420 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.384442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.384457 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.388181 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.399609 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402549 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402595 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-kubelet\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402618 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-conf-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402642 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n4v8\" (UniqueName: \"kubernetes.io/projected/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-kube-api-access-2n4v8\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402670 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qh4j\" (UniqueName: \"kubernetes.io/projected/fdb45e62-a601-42e3-988e-4834834ecb6b-kube-api-access-4qh4j\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402735 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-binary-copy\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402755 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-k8s-cni-cncf-io\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402774 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-hostroot\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402791 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-os-release\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402810 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402842 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-netns\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402859 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-cnibin\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402875 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-cnibin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402917 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-cni-binary-copy\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402944 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-bin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402963 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-socket-dir-parent\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.402979 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-multus\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403002 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-proxy-tls\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403019 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403041 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-daemon-config\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403146 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-system-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403188 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-system-cni-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403205 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-os-release\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403246 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjcf\" (UniqueName: \"kubernetes.io/projected/aff2b22c-b87d-47c9-b5a0-6e5772022488-kube-api-access-6sjcf\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403271 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-rootfs\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403328 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-multus-certs\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403373 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-etc-kubernetes\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.403411 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-mcd-auth-proxy-config\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.416084 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.439100 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.452715 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.463084 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.479404 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.487633 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.487670 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.487680 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.487697 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.487709 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.495047 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504420 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504508 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-system-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504533 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-system-cni-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504553 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-os-release\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504624 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-os-release\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.504635 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:24.504599904 +0000 UTC m=+19.984902942 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504669 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-system-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504712 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504784 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjcf\" (UniqueName: \"kubernetes.io/projected/aff2b22c-b87d-47c9-b5a0-6e5772022488-kube-api-access-6sjcf\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.504815 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504817 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-rootfs\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.504832 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.504844 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504845 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-multus-certs\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504872 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.504906 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:24.504870403 +0000 UTC m=+19.985173441 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504922 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-etc-kubernetes\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504946 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504964 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-mcd-auth-proxy-config\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.504983 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505001 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-kubelet\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505018 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-conf-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505036 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n4v8\" (UniqueName: \"kubernetes.io/projected/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-kube-api-access-2n4v8\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505056 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qh4j\" (UniqueName: \"kubernetes.io/projected/fdb45e62-a601-42e3-988e-4834834ecb6b-kube-api-access-4qh4j\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505087 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-hostroot\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505110 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-binary-copy\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505133 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-k8s-cni-cncf-io\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505139 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505166 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-os-release\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505169 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:24.505162491 +0000 UTC m=+19.985465529 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505192 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-system-cni-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505207 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505237 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505262 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-netns\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505282 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-cnibin\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505304 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-cnibin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505325 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-cni-binary-copy\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505350 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-bin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505358 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-rootfs\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505373 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-proxy-tls\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505389 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-multus-certs\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505403 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-socket-dir-parent\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505434 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-multus\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505446 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505456 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505459 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505482 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-daemon-config\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505482 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-etc-kubernetes\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505696 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-netns\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505734 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-cnibin\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.505793 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-cnibin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506071 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506211 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-hostroot\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506326 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-kubelet\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505465 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.505665 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506446 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-cni-binary-copy\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506439 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-conf-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506554 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-multus\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506589 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-socket-dir-parent\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506580 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-run-k8s-cni-cncf-io\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506622 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-host-var-lib-cni-bin\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506753 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fdb45e62-a601-42e3-988e-4834834ecb6b-cni-binary-copy\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.506754 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:24.506715404 +0000 UTC m=+19.987018612 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506754 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-os-release\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: E0122 00:07:23.506802 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:24.506795216 +0000 UTC m=+19.987098254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.506934 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fdb45e62-a601-42e3-988e-4834834ecb6b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.507118 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-cni-dir\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.507253 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/aff2b22c-b87d-47c9-b5a0-6e5772022488-multus-daemon-config\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.507743 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-mcd-auth-proxy-config\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.508408 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.513794 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-proxy-tls\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.524125 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qh4j\" (UniqueName: \"kubernetes.io/projected/fdb45e62-a601-42e3-988e-4834834ecb6b-kube-api-access-4qh4j\") pod \"multus-additional-cni-plugins-bv9wt\" (UID: \"fdb45e62-a601-42e3-988e-4834834ecb6b\") " pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.525839 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.529191 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjcf\" (UniqueName: \"kubernetes.io/projected/aff2b22c-b87d-47c9-b5a0-6e5772022488-kube-api-access-6sjcf\") pod \"multus-vfgd4\" (UID: \"aff2b22c-b87d-47c9-b5a0-6e5772022488\") " pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.529921 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n4v8\" (UniqueName: \"kubernetes.io/projected/8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1-kube-api-access-2n4v8\") pod \"machine-config-daemon-mrfxg\" (UID: \"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\") " pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.537133 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.548824 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.557018 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.568134 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.580334 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.590525 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.590839 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.590996 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.591071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.591149 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.592003 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.616451 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vfgd4" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.624944 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.642217 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.653645 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w4dww"] Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.654788 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.658609 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.658674 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.658830 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.659436 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.659851 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.661718 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.664500 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.671420 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.685400 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.698987 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.699040 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.699058 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.699080 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.699101 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: W0122 00:07:23.700244 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaff2b22c_b87d_47c9_b5a0_6e5772022488.slice/crio-d99d413873e5e4fe888359226e6c8eee98b6d0135a2805236a1334ec7de4a78a WatchSource:0}: Error finding container d99d413873e5e4fe888359226e6c8eee98b6d0135a2805236a1334ec7de4a78a: Status 404 returned error can't find the container with id d99d413873e5e4fe888359226e6c8eee98b6d0135a2805236a1334ec7de4a78a Jan 22 00:07:23 crc kubenswrapper[4800]: W0122 00:07:23.702904 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdb45e62_a601_42e3_988e_4834834ecb6b.slice/crio-8d9f39f4226dbf1cf1a1d52dbd118b8de0bc8ab602d14121896693c87e1b2347 WatchSource:0}: Error finding container 8d9f39f4226dbf1cf1a1d52dbd118b8de0bc8ab602d14121896693c87e1b2347: Status 404 returned error can't find the container with id 8d9f39f4226dbf1cf1a1d52dbd118b8de0bc8ab602d14121896693c87e1b2347 Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707667 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707709 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707727 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707749 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707767 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707783 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707800 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707819 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gx9m\" (UniqueName: \"kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707847 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707865 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707898 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707934 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707951 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707969 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.707985 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708004 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708024 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708048 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708082 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708105 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.708473 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.725994 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.742607 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.759958 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.771221 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 01:24:34.540687087 +0000 UTC Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.777987 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.793488 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.803454 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.803517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.803529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.803553 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.803566 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810271 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810308 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810326 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810343 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810367 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810380 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810399 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810419 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810435 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810451 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810468 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810504 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gx9m\" (UniqueName: \"kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810519 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810534 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810549 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810588 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810604 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810623 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810640 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810655 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.810743 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.811702 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812206 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812276 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812245 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812274 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812222 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812397 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812462 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812468 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812444 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812468 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812502 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812520 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812507 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812464 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812748 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.813004 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.812731 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.815038 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.827192 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.831433 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gx9m\" (UniqueName: \"kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m\") pod \"ovnkube-node-w4dww\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.843526 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.857870 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.907330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.907375 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.907389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.907410 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.907423 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:23Z","lastTransitionTime":"2026-01-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.958044 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerStarted","Data":"7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.958133 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerStarted","Data":"d99d413873e5e4fe888359226e6c8eee98b6d0135a2805236a1334ec7de4a78a"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.959743 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerStarted","Data":"8d9f39f4226dbf1cf1a1d52dbd118b8de0bc8ab602d14121896693c87e1b2347"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.962795 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fp7s5" event={"ID":"b37dab4b-08e1-4daa-86d9-561432a82428","Type":"ContainerStarted","Data":"6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.962841 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fp7s5" event={"ID":"b37dab4b-08e1-4daa-86d9-561432a82428","Type":"ContainerStarted","Data":"ea54dcf167367e1bbee0cbc4ce74ff2ff0755bc86044639a01068ba52212a864"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.963789 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ab311420ac3b38ff4bac59dce09846554d3c55521e90168b65c7ed74e3ecb471"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.966847 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.969476 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.969540 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.969561 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"10a08d74f9cace0b37f170da8fcfb383584ffc3a20fcc89dae715698a97c9541"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.971965 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.972037 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1e0fe8e411d6dbc649cfbf26277772d67e8a16f3c436e499e04b55c501618049"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.977662 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.977710 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.977728 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"50ec02560f9dec719951fdb8cfb2f2732ca599ae286143f8a1e61251a023cf23"} Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.977776 4800 scope.go:117] "RemoveContainer" containerID="79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003" Jan 22 00:07:23 crc kubenswrapper[4800]: I0122 00:07:23.992036 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:23Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.011599 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.011645 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.011659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.011681 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.011694 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.026721 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.057745 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.075389 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.098704 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117019 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117075 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117095 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117108 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.117856 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.135194 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.150410 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.164059 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.178034 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.192656 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.210106 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.220190 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.220306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.220367 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.220451 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.220516 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.277985 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.318250 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.324971 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.325018 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.325031 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.325052 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.325066 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.342382 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.385143 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.405995 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.424134 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.428265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.428286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.428296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.428313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.428325 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.437383 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.447825 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.462121 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.485210 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.507681 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.518618 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.518743 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.518779 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.518805 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.518880 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519092 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519133 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519164 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519183 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519167 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:26.519146531 +0000 UTC m=+21.999449589 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519313 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:26.519294495 +0000 UTC m=+21.999597533 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519326 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:26.519320675 +0000 UTC m=+21.999623713 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519362 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519382 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:26.519377567 +0000 UTC m=+21.999680605 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519423 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519434 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519442 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.519459 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:26.519454079 +0000 UTC m=+21.999757117 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.531448 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.531503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.531518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.531543 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.531559 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.549432 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.634315 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.634372 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.634391 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.634419 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.634436 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.667772 4800 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.741295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.741683 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.741695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.741713 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.741725 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.771956 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:44:09.468840687 +0000 UTC Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.817749 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.817801 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.817979 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.818019 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.818101 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:24 crc kubenswrapper[4800]: E0122 00:07:24.818191 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.822199 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.823149 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.824403 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.825197 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.826609 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.827255 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.827863 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.829489 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.830182 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.831355 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.831965 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.833088 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.833614 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.833785 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.834176 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.835229 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.835942 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.837160 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.837634 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.838197 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.839235 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.839670 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.840748 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.841184 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.842492 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.843760 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844090 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844179 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844211 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844237 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.844540 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.845640 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.846147 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.847322 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.847877 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.848844 4800 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.848994 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.850703 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.851371 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.852234 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.852797 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.854368 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.855090 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.859906 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.860673 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.861763 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.862230 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.863251 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.863928 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.864924 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.865400 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.866620 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.867219 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.874573 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.875146 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.876274 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.877529 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.878672 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.879413 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.879917 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.884339 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.903873 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.920400 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.940633 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.947144 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.947212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.947225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.947242 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.947272 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:24Z","lastTransitionTime":"2026-01-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.957547 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.972560 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.982006 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72" exitCode=0 Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.982107 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.983795 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" exitCode=0 Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.983875 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.983932 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"76fe4bd599e7342a097e8909937e6c95b31444d2c51f11efe297a01f00745da7"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.987155 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.989709 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174"} Jan 22 00:07:24 crc kubenswrapper[4800]: I0122 00:07:24.990171 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.001959 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.040359 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.054358 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.055113 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.055135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.055154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.055166 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.059214 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.085271 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.103127 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.130115 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.154246 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.157696 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.157739 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.157750 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.157769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.157809 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.265164 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.270652 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.270682 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.270690 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.270703 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.270713 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.289661 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.308212 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.323159 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.361161 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.373587 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.373627 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.373636 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.373651 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.373662 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.388031 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.427552 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.476153 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.476754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.476865 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.476955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.477045 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.477113 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.510489 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.580330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.580376 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.580385 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.580404 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.580416 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.683677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.684046 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.684057 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.684083 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.684096 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.772763 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 18:29:20.555983043 +0000 UTC Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.787241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.787314 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.787337 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.787371 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.787393 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.897722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.898093 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.898168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.898253 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.898315 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:25Z","lastTransitionTime":"2026-01-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.997866 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16" exitCode=0 Jan 22 00:07:25 crc kubenswrapper[4800]: I0122 00:07:25.997950 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.002540 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.002599 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.002617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.002666 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.002683 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005178 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005335 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005404 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005462 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005540 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.005602 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.026175 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.045169 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.061984 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.084412 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.098144 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.105972 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.106017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.106032 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.106052 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.106064 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.118664 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.134221 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.149013 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.163624 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.176518 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.191437 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.205173 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.209582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.209657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.209676 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.209706 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.209726 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.313286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.313336 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.313351 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.313372 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.313385 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.356621 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7tj9m"] Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.357073 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.360089 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.360321 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.360632 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.362921 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.376625 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.406720 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.415973 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.416168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.416250 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.416343 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.416431 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.421402 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.439981 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.444073 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc6ll\" (UniqueName: \"kubernetes.io/projected/7e29e23b-f99c-4c89-8126-9493b6421e99-kube-api-access-rc6ll\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.444240 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e29e23b-f99c-4c89-8126-9493b6421e99-host\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.444404 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e29e23b-f99c-4c89-8126-9493b6421e99-serviceca\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.460243 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.485071 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.505580 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.519507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.519552 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.519561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.519579 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.519591 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.523639 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.540773 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545038 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545222 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545327 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:30.54528145 +0000 UTC m=+26.025584518 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545392 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545424 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545447 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545511 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545529 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:30.545501156 +0000 UTC m=+26.025804234 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545422 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.545565 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:30.545550037 +0000 UTC m=+26.025853075 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545719 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e29e23b-f99c-4c89-8126-9493b6421e99-serviceca\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545828 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc6ll\" (UniqueName: \"kubernetes.io/projected/7e29e23b-f99c-4c89-8126-9493b6421e99-kube-api-access-rc6ll\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545869 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545952 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e29e23b-f99c-4c89-8126-9493b6421e99-host\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.545975 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546112 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546185 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:30.546169515 +0000 UTC m=+26.026472583 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546189 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.546099 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e29e23b-f99c-4c89-8126-9493b6421e99-host\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546217 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546231 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.546305 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:30.546283718 +0000 UTC m=+26.026586756 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.548679 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e29e23b-f99c-4c89-8126-9493b6421e99-serviceca\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.563116 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.574322 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.578844 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc6ll\" (UniqueName: \"kubernetes.io/projected/7e29e23b-f99c-4c89-8126-9493b6421e99-kube-api-access-rc6ll\") pod \"node-ca-7tj9m\" (UID: \"7e29e23b-f99c-4c89-8126-9493b6421e99\") " pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.582065 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.586939 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.606605 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.622977 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.623016 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.623026 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.623043 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.623056 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.624249 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.644567 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.666580 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.706441 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7tj9m" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.712593 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.726471 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.726765 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.727058 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.727802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.728027 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: W0122 00:07:26.736460 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e29e23b_f99c_4c89_8126_9493b6421e99.slice/crio-e4d84fb1c8f92b3bbfe6b9cb32dfb801b1211a608bbb7bad0b50582d9966f866 WatchSource:0}: Error finding container e4d84fb1c8f92b3bbfe6b9cb32dfb801b1211a608bbb7bad0b50582d9966f866: Status 404 returned error can't find the container with id e4d84fb1c8f92b3bbfe6b9cb32dfb801b1211a608bbb7bad0b50582d9966f866 Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.765488 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.773614 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 13:11:44.176247587 +0000 UTC Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.797629 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.817474 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.817626 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.818027 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.818080 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.818223 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:26 crc kubenswrapper[4800]: E0122 00:07:26.818277 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.833633 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.833695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.833707 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.833748 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.833765 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.835429 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.869265 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.907366 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.936638 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.936705 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.936725 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.936754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.936772 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:26Z","lastTransitionTime":"2026-01-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.947832 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:26 crc kubenswrapper[4800]: I0122 00:07:26.988591 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:26Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.009531 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7tj9m" event={"ID":"7e29e23b-f99c-4c89-8126-9493b6421e99","Type":"ContainerStarted","Data":"e4d84fb1c8f92b3bbfe6b9cb32dfb801b1211a608bbb7bad0b50582d9966f866"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.011173 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.013377 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc" exitCode=0 Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.013549 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.026840 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.039525 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.039564 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.039582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.039601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.039610 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: E0122 00:07:27.043007 4800 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.088340 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.127366 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.131671 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146249 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146323 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146346 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146359 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.146654 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.175658 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.190087 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.227017 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.249938 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.250027 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.250044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.250071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.250089 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.269052 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.305456 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.350207 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.352398 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.352428 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.352439 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.352459 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.352471 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.386593 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.427582 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.455545 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.455580 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.455592 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.455612 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.455629 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.470835 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.508722 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.553104 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.560010 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.560040 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.560050 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.560069 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.560081 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.595152 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.636625 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.662916 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.662991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.663010 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.663045 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.663063 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.669532 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.714298 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.754489 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.767681 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.767752 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.767779 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.767814 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.767836 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.773969 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 18:26:07.119785789 +0000 UTC Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.803427 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.830217 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:27Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.872030 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.872120 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.872147 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.872181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.872203 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.975734 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.975804 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.975822 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.975848 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:27 crc kubenswrapper[4800]: I0122 00:07:27.975866 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:27Z","lastTransitionTime":"2026-01-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.022639 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8" exitCode=0 Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.022743 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.024939 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7tj9m" event={"ID":"7e29e23b-f99c-4c89-8126-9493b6421e99","Type":"ContainerStarted","Data":"7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.043381 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.067499 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.080854 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.080922 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.080933 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.080954 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.080967 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.083787 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.098988 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.113544 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.131767 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.145967 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.163847 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.201612 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.201657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.201669 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.201690 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.201703 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.214526 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.248367 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.271731 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.303462 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.303513 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.303528 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.303550 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.303567 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.309039 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.348967 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.391012 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.406201 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.406237 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.406247 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.406266 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.406278 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.430469 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.474533 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.507020 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.508930 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.508991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.509005 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.509065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.509092 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.549248 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.588860 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.612308 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.612346 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.612378 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.612396 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.612408 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.628570 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.671018 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.716456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.716503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.716514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.716537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.716554 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.722400 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.749057 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.774712 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 06:01:14.246316313 +0000 UTC Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.799377 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.818101 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.818146 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.818157 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:28 crc kubenswrapper[4800]: E0122 00:07:28.818359 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:28 crc kubenswrapper[4800]: E0122 00:07:28.818510 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:28 crc kubenswrapper[4800]: E0122 00:07:28.818691 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.820728 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.820774 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.820789 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.820811 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.820827 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.832167 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.872336 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.914869 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.923493 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.923586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.923617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.923650 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.923671 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:28Z","lastTransitionTime":"2026-01-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.956491 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:28 crc kubenswrapper[4800]: I0122 00:07:28.987642 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:28Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.026683 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.027030 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.027128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.027194 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.027251 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.030302 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.031634 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.034721 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22" exitCode=0 Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.034826 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.069376 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.116142 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.130245 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.130287 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.130295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.130311 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.130321 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.148668 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.189308 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.230367 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.233421 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.233473 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.233484 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.233504 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.233515 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.267502 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.311166 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.335792 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.335871 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.335906 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.335929 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.335943 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.350740 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.391366 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.428100 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.439432 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.439483 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.439501 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.439531 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.439552 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.467114 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.506877 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.542159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.542226 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.542239 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.542262 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.542276 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.551462 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.592466 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.629186 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:29Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.645808 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.645861 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.645879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.645929 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.645944 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.749155 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.749206 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.749222 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.749253 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.749266 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.775868 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 10:22:56.949662464 +0000 UTC Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.852050 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.852169 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.852195 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.852243 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.852277 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.955537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.955578 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.955589 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.955605 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:29 crc kubenswrapper[4800]: I0122 00:07:29.955616 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:29Z","lastTransitionTime":"2026-01-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.041623 4800 generic.go:334] "Generic (PLEG): container finished" podID="fdb45e62-a601-42e3-988e-4834834ecb6b" containerID="9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786" exitCode=0 Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.041670 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerDied","Data":"9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.057342 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.057413 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.057428 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.057448 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.057461 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.062548 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.082232 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.097692 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.108690 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.124509 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.147807 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.159969 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.160012 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.160022 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.160037 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.160050 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.166983 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.182651 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.201390 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.213614 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.226779 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.247780 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.261864 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.262777 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.262819 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.262831 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.262847 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.262859 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.275035 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.290582 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:30Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.367095 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.367136 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.367147 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.367165 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.367176 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.470332 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.470383 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.470397 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.470417 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.470428 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.573292 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.573347 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.573362 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.573388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.573407 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.589958 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590130 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.590101031 +0000 UTC m=+34.070404079 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.590189 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.590241 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.590301 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590398 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590452 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.59044142 +0000 UTC m=+34.070744468 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590453 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590476 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590492 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590530 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.590518472 +0000 UTC m=+34.070821520 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.590552 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590638 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590725 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590766 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590809 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.590752 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.590728268 +0000 UTC m=+34.071031306 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.591177 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.59115313 +0000 UTC m=+34.071456168 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.676641 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.676685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.676697 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.676712 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.676722 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.776561 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 22:25:52.466214332 +0000 UTC Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.780267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.780317 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.780330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.780354 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.780366 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.817646 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.817831 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.817931 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.817663 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.818121 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:30 crc kubenswrapper[4800]: E0122 00:07:30.818212 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.882730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.882771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.882781 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.882795 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.882809 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.986455 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.986527 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.986547 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.986578 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:30 crc kubenswrapper[4800]: I0122 00:07:30.986603 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:30Z","lastTransitionTime":"2026-01-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.054673 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.055275 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.055386 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.055399 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.061009 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" event={"ID":"fdb45e62-a601-42e3-988e-4834834ecb6b","Type":"ContainerStarted","Data":"23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.075093 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.086711 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.089882 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.089973 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.089991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.090019 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.090038 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.090213 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.098978 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.115719 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.132753 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.148792 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.166379 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.182855 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.193482 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.193530 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.193545 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.193565 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.193581 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.198509 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.218777 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.235998 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.251972 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.266146 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.278234 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.293517 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.295769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.295851 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.295872 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.295941 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.295971 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.323746 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.340605 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.354316 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.372872 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.384583 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.398781 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.398989 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.399060 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.398975 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.399128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.399311 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.412503 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.430752 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.454377 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.472665 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.488368 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.502252 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.502287 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.502296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.502312 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.502324 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.506236 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.527265 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.546068 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.564582 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.580425 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:31Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.604802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.604877 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.604937 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.604971 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.604992 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.708224 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.708604 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.708707 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.708835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.708971 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.777085 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 20:27:23.908121192 +0000 UTC Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.811727 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.811783 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.811797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.811823 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.811842 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.914611 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.914646 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.914655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.914672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:31 crc kubenswrapper[4800]: I0122 00:07:31.914681 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:31Z","lastTransitionTime":"2026-01-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.016932 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.016976 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.016991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.017008 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.017020 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.120871 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.120949 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.120961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.120982 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.120995 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.224218 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.224667 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.224817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.225024 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.225195 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.328666 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.328713 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.328723 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.328741 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.328752 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.432091 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.432128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.432142 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.432168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.432181 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.535556 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.535628 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.535655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.535683 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.535704 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.638269 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.638340 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.638360 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.638387 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.638408 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.747774 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.747850 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.747870 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.747925 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.747942 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.778090 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 02:51:58.355317764 +0000 UTC Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.817720 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.817720 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:32 crc kubenswrapper[4800]: E0122 00:07:32.817934 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:32 crc kubenswrapper[4800]: E0122 00:07:32.818043 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.817720 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:32 crc kubenswrapper[4800]: E0122 00:07:32.818175 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.850061 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.850361 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.850492 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.850672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.850765 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.954442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.954478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.954489 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.954507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:32 crc kubenswrapper[4800]: I0122 00:07:32.954518 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:32Z","lastTransitionTime":"2026-01-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.057648 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.057705 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.057720 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.057738 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.057752 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.069839 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/0.log" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.073226 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1" exitCode=1 Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.073273 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.074126 4800 scope.go:117] "RemoveContainer" containerID="c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.088750 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.105111 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.120240 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.135637 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.148451 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161848 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161903 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161915 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161928 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161907 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.161937 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.178905 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.203323 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:32Z\\\",\\\"message\\\":\\\"rsions/factory.go:140\\\\nI0122 00:07:32.414337 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:32.414364 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:32.414431 6099 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:32.414498 6099 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 00:07:32.414510 6099 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 00:07:32.414561 6099 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 00:07:32.414577 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:32.414587 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:32.414592 6099 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:32.414600 6099 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:32.414609 6099 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 00:07:32.415212 6099 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 00:07:32.415263 6099 factory.go:656] Stopping watch factory\\\\nI0122 00:07:32.415303 6099 ovnkube.go:599] Stopped ovnkube\\\\nI0122 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.217849 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.233123 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.248201 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.264547 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.264660 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.264675 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.264691 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.264701 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.265297 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.277933 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.293810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.293859 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.293868 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.293907 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.293919 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.309682 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.309992 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.313681 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.313721 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.313731 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.313748 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.313759 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.325779 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.326649 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.330098 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.330128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.330138 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.330154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.330165 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.343402 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.347474 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.347514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.347524 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.347545 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.347559 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.364777 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.373221 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.373326 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.373341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.373368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.373389 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.401120 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:33Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:33 crc kubenswrapper[4800]: E0122 00:07:33.401250 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.403322 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.403361 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.403376 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.403394 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.403405 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.506116 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.506160 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.506219 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.506236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.506247 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.608579 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.608617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.608630 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.608648 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.608660 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.711425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.711465 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.711474 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.711489 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.711500 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.778389 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 12:27:57.408868845 +0000 UTC Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.814524 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.814598 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.814622 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.814657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.814681 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.917610 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.917680 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.917735 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.917769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:33 crc kubenswrapper[4800]: I0122 00:07:33.917788 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:33Z","lastTransitionTime":"2026-01-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.020685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.020741 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.020751 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.020771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.020784 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.081076 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/0.log" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.085524 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.086079 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.117235 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.125377 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.125442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.125456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.125485 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.125503 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.136495 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.152021 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.166098 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.183431 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.196430 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.205377 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.216822 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.228146 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.228203 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.228224 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.228251 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.228269 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.232169 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.246726 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.260610 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.272003 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.331481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.331551 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.331568 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.331595 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.331614 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.434752 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.434828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.434839 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.434858 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.434870 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.468854 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.503329 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:32Z\\\",\\\"message\\\":\\\"rsions/factory.go:140\\\\nI0122 00:07:32.414337 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:32.414364 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:32.414431 6099 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:32.414498 6099 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 00:07:32.414510 6099 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 00:07:32.414561 6099 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 00:07:32.414577 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:32.414587 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:32.414592 6099 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:32.414600 6099 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:32.414609 6099 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 00:07:32.415212 6099 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 00:07:32.415263 6099 factory.go:656] Stopping watch factory\\\\nI0122 00:07:32.415303 6099 ovnkube.go:599] Stopped ovnkube\\\\nI0122 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.525114 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.538116 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.538182 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.538236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.538268 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.538289 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.641471 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.641551 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.641582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.641618 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.641641 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.745451 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.745536 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.745561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.745600 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.745622 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.779589 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 19:06:03.794243599 +0000 UTC Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.817614 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.817700 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.817729 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:34 crc kubenswrapper[4800]: E0122 00:07:34.818075 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:34 crc kubenswrapper[4800]: E0122 00:07:34.818178 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:34 crc kubenswrapper[4800]: E0122 00:07:34.818287 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.840165 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.849228 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.849280 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.849295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.849313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.849327 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.862591 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.881725 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.905044 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.924304 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.947327 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.952774 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.952811 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.952821 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.952838 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.952849 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:34Z","lastTransitionTime":"2026-01-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.976653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:32Z\\\",\\\"message\\\":\\\"rsions/factory.go:140\\\\nI0122 00:07:32.414337 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:32.414364 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:32.414431 6099 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:32.414498 6099 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 00:07:32.414510 6099 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 00:07:32.414561 6099 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 00:07:32.414577 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:32.414587 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:32.414592 6099 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:32.414600 6099 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:32.414609 6099 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 00:07:32.415212 6099 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 00:07:32.415263 6099 factory.go:656] Stopping watch factory\\\\nI0122 00:07:32.415303 6099 ovnkube.go:599] Stopped ovnkube\\\\nI0122 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:34 crc kubenswrapper[4800]: I0122 00:07:34.993826 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:34Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.019419 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.036003 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.050763 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.056169 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.056203 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.056217 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.056235 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.056246 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.066876 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.088214 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.091144 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/1.log" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.092006 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/0.log" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.094949 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d" exitCode=1 Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.095000 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.095042 4800 scope.go:117] "RemoveContainer" containerID="c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.095953 4800 scope.go:117] "RemoveContainer" containerID="6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d" Jan 22 00:07:35 crc kubenswrapper[4800]: E0122 00:07:35.096163 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.103817 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.119153 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.132771 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.159191 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.159245 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.159255 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.159276 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.159289 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.167668 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.183770 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.197103 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.212966 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.229106 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.249153 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.262288 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.262352 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.262371 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.262398 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.262414 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.265666 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.278555 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.294676 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.316827 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:32Z\\\",\\\"message\\\":\\\"rsions/factory.go:140\\\\nI0122 00:07:32.414337 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:32.414364 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:32.414431 6099 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:32.414498 6099 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 00:07:32.414510 6099 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 00:07:32.414561 6099 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 00:07:32.414577 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:32.414587 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:32.414592 6099 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:32.414600 6099 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:32.414609 6099 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 00:07:32.415212 6099 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 00:07:32.415263 6099 factory.go:656] Stopping watch factory\\\\nI0122 00:07:32.415303 6099 ovnkube.go:599] Stopped ovnkube\\\\nI0122 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.330542 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.349257 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.364470 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.366316 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.366349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.366357 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.366372 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.366383 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.378964 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.468637 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.468678 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.468687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.468701 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.468712 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.571963 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.572021 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.572039 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.572069 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.572091 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.675194 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.675237 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.675246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.675260 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.675269 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.778305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.778391 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.778420 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.778451 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.778475 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.780417 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 03:23:58.048092862 +0000 UTC Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.882229 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.882324 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.882340 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.882358 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.882368 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.902846 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.918782 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.935453 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.949449 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.964131 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.984161 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99c6f91c5f8c1f2507bc6537f3fd5b48d24cddf73eac4234f6784cfa49841b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:32Z\\\",\\\"message\\\":\\\"rsions/factory.go:140\\\\nI0122 00:07:32.414337 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:32.414364 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:32.414431 6099 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:32.414498 6099 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 00:07:32.414510 6099 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 00:07:32.414561 6099 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 00:07:32.414577 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:32.414587 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:32.414592 6099 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:32.414600 6099 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:32.414609 6099 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0122 00:07:32.415212 6099 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 00:07:32.415263 6099 factory.go:656] Stopping watch factory\\\\nI0122 00:07:32.415303 6099 ovnkube.go:599] Stopped ovnkube\\\\nI0122 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.985294 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.985344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.985364 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.985388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.985403 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:35Z","lastTransitionTime":"2026-01-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:35 crc kubenswrapper[4800]: I0122 00:07:35.998777 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:35Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.011965 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.023335 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.040300 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.062699 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.079724 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.088803 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.088863 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.088900 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.088926 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.088941 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.098465 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.100522 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/1.log" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.103964 4800 scope.go:117] "RemoveContainer" containerID="6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d" Jan 22 00:07:36 crc kubenswrapper[4800]: E0122 00:07:36.104227 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.109393 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm"] Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.109828 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: W0122 00:07:36.111407 4800 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Jan 22 00:07:36 crc kubenswrapper[4800]: E0122 00:07:36.111494 4800 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.111937 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.119974 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.138291 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.153829 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.167098 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.180678 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.192431 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.192525 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.192546 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.192581 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.192603 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.196792 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.211921 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.225653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.254155 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.254221 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.254273 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.254308 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2bm5\" (UniqueName: \"kubernetes.io/projected/9abc3993-c3c9-4b93-a963-13dee17b8cff-kube-api-access-b2bm5\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.258333 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.276205 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295109 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295187 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295210 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295228 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.295869 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.312581 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.331438 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.344643 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.355778 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.355845 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.355944 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.355988 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2bm5\" (UniqueName: \"kubernetes.io/projected/9abc3993-c3c9-4b93-a963-13dee17b8cff-kube-api-access-b2bm5\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.356942 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.357280 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.360173 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.364673 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9abc3993-c3c9-4b93-a963-13dee17b8cff-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.377061 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.384778 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2bm5\" (UniqueName: \"kubernetes.io/projected/9abc3993-c3c9-4b93-a963-13dee17b8cff-kube-api-access-b2bm5\") pod \"ovnkube-control-plane-749d76644c-7zgrm\" (UID: \"9abc3993-c3c9-4b93-a963-13dee17b8cff\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.391790 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.397328 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.397365 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.397374 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.397389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.397399 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.414827 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.427215 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:36Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.499763 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.499799 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.499811 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.499825 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.499836 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.602633 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.602729 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.602752 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.602782 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.602805 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.706135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.706206 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.706232 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.706265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.706290 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.781010 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 23:07:29.679817392 +0000 UTC Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.809942 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.810349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.810586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.810768 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.810982 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.818053 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.818179 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:36 crc kubenswrapper[4800]: E0122 00:07:36.818232 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.818288 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:36 crc kubenswrapper[4800]: E0122 00:07:36.818458 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:36 crc kubenswrapper[4800]: E0122 00:07:36.818573 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.914566 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.914948 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.915284 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.915426 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:36 crc kubenswrapper[4800]: I0122 00:07:36.915528 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:36Z","lastTransitionTime":"2026-01-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.018344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.018425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.018442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.018460 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.018475 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.121235 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.121307 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.121344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.121364 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.121377 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.224753 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.225106 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.225211 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.225295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.225361 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.329959 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.330479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.330852 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.331091 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.331222 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.424468 4800 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" secret="" err="failed to sync secret cache: timed out waiting for the condition" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.424571 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.434286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.434325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.434335 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.434353 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.434364 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: W0122 00:07:37.443577 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9abc3993_c3c9_4b93_a963_13dee17b8cff.slice/crio-43cf767ad2b9a69f8f75de0978eb8e6aacac5e91ca15663c00a094b368d1b25b WatchSource:0}: Error finding container 43cf767ad2b9a69f8f75de0978eb8e6aacac5e91ca15663c00a094b368d1b25b: Status 404 returned error can't find the container with id 43cf767ad2b9a69f8f75de0978eb8e6aacac5e91ca15663c00a094b368d1b25b Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.537275 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.537820 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.537833 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.537852 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.537867 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.594635 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-svkb9"] Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.595375 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: E0122 00:07:37.595466 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.614053 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.627167 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.641220 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.641293 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.641308 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.641327 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.641361 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.642830 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.667065 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.667080 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.669775 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxlxs\" (UniqueName: \"kubernetes.io/projected/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-kube-api-access-nxlxs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.669879 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.683680 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.702291 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.717297 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.738873 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.743663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.743722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.743737 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.743763 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.743781 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.755075 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.766292 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.770511 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxlxs\" (UniqueName: \"kubernetes.io/projected/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-kube-api-access-nxlxs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.770578 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: E0122 00:07:37.770684 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:37 crc kubenswrapper[4800]: E0122 00:07:37.770737 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:38.270721647 +0000 UTC m=+33.751024695 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.781783 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 11:31:14.885494178 +0000 UTC Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.792569 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.798918 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxlxs\" (UniqueName: \"kubernetes.io/projected/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-kube-api-access-nxlxs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.812045 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.828147 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.843591 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.846646 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.846771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.846783 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.846800 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.846810 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.861694 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.880905 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.896838 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:37Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.949776 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.949828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.949840 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.949859 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:37 crc kubenswrapper[4800]: I0122 00:07:37.950389 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:37Z","lastTransitionTime":"2026-01-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.053282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.053324 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.053339 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.053355 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.053366 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.114058 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" event={"ID":"9abc3993-c3c9-4b93-a963-13dee17b8cff","Type":"ContainerStarted","Data":"2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.114115 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" event={"ID":"9abc3993-c3c9-4b93-a963-13dee17b8cff","Type":"ContainerStarted","Data":"1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.114129 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" event={"ID":"9abc3993-c3c9-4b93-a963-13dee17b8cff","Type":"ContainerStarted","Data":"43cf767ad2b9a69f8f75de0978eb8e6aacac5e91ca15663c00a094b368d1b25b"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.127500 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.139984 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.150197 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.155474 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.155529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.155543 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.155560 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.155571 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.164668 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.176472 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.185924 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.197762 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.214001 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.225749 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.238648 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.250584 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.258650 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.258686 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.258695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.258713 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.258723 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.262480 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.274576 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.275052 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.275194 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.275256 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:39.275239808 +0000 UTC m=+34.755542846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.285803 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.296215 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.315189 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.327626 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:38Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.362368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.362425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.362436 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.362453 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.362463 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.465990 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.466042 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.466052 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.466069 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.466082 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.569624 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.569679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.569694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.569710 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.569720 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.672730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.672816 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.672831 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.672845 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.672856 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.679341 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.679501 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.679611 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679700 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679729 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:07:54.67967801 +0000 UTC m=+50.159981098 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679777 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679793 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:54.679770993 +0000 UTC m=+50.160074161 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679801 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679823 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.679849 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.679882 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:54.679860245 +0000 UTC m=+50.160163323 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.679962 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680088 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680166 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:54.680145913 +0000 UTC m=+50.160448971 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680196 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680242 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680269 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.680360 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:54.680337208 +0000 UTC m=+50.160640296 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.775678 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.775722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.775734 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.775748 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.775758 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.782079 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 00:46:48.319190536 +0000 UTC Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.817684 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.817760 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.817799 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.817808 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.817944 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.818092 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.818205 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:38 crc kubenswrapper[4800]: E0122 00:07:38.818284 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.878452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.878490 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.878500 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.878514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.878522 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.981984 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.982032 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.982042 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.982061 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:38 crc kubenswrapper[4800]: I0122 00:07:38.982070 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:38Z","lastTransitionTime":"2026-01-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.085420 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.085506 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.085530 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.085562 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.085585 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.188510 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.188563 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.188581 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.188601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.188617 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.286584 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:39 crc kubenswrapper[4800]: E0122 00:07:39.286771 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:39 crc kubenswrapper[4800]: E0122 00:07:39.286840 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:41.286821861 +0000 UTC m=+36.767124899 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.291789 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.291873 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.291938 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.291979 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.292003 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.394542 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.394650 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.394660 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.394674 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.394683 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.497209 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.497279 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.497288 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.497304 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.497315 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.599909 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.599953 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.599965 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.599980 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.599990 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.701864 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.701917 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.701928 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.701945 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.701959 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.782839 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 02:19:19.825020035 +0000 UTC Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.804329 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.804392 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.804411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.804433 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.804451 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.906678 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.906759 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.906775 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.906792 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:39 crc kubenswrapper[4800]: I0122 00:07:39.906803 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:39Z","lastTransitionTime":"2026-01-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.009097 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.009139 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.009151 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.009168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.009179 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.112773 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.112839 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.112864 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.112930 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.112958 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.215938 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.216014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.216036 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.216063 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.216082 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.318951 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.318991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.319002 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.319017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.319028 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.421757 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.421804 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.421812 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.421830 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.421839 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.524606 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.524649 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.524660 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.524676 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.524688 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.627979 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.628050 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.628071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.628100 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.628121 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.731843 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.731945 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.731964 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.731994 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.732012 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.783339 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 19:10:17.372669144 +0000 UTC Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.818187 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.818259 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.818267 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:40 crc kubenswrapper[4800]: E0122 00:07:40.818427 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.818530 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:40 crc kubenswrapper[4800]: E0122 00:07:40.818618 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:40 crc kubenswrapper[4800]: E0122 00:07:40.818729 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:40 crc kubenswrapper[4800]: E0122 00:07:40.818821 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.834797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.834864 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.834975 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.835026 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.835061 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.938779 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.938831 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.938841 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.938858 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:40 crc kubenswrapper[4800]: I0122 00:07:40.938868 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:40Z","lastTransitionTime":"2026-01-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.042272 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.042305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.042314 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.042329 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.042341 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.145308 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.145348 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.145357 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.145371 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.145381 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.247878 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.247961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.247974 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.247997 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.248013 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.309963 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:41 crc kubenswrapper[4800]: E0122 00:07:41.310194 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:41 crc kubenswrapper[4800]: E0122 00:07:41.310298 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:45.310277756 +0000 UTC m=+40.790580794 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.350313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.350354 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.350364 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.350379 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.350390 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.453864 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.453962 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.453982 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.454010 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.454029 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.557509 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.557621 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.557644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.557688 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.557712 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.660483 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.660554 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.660571 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.660595 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.660608 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.763475 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.763530 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.763547 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.763607 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.763628 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.783922 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 19:19:37.220966175 +0000 UTC Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.866957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.867026 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.867047 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.867074 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.867094 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.970332 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.970395 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.970411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.970438 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:41 crc kubenswrapper[4800]: I0122 00:07:41.970458 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:41Z","lastTransitionTime":"2026-01-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.074186 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.074238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.074247 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.074263 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.074273 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.177132 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.177208 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.177230 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.177261 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.177284 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.280654 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.280722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.280735 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.280756 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.280770 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.383618 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.383687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.383710 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.383741 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.383765 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.486406 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.486443 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.486452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.486466 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.486475 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.590003 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.590056 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.590065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.590078 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.590089 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.693555 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.693620 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.693640 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.693663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.693680 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.785043 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 01:11:13.722356637 +0000 UTC Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.796529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.796574 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.796583 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.796612 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.796626 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.818017 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.818126 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.818228 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:42 crc kubenswrapper[4800]: E0122 00:07:42.818217 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:42 crc kubenswrapper[4800]: E0122 00:07:42.818400 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.818477 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:42 crc kubenswrapper[4800]: E0122 00:07:42.818531 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:42 crc kubenswrapper[4800]: E0122 00:07:42.818678 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.899507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.899567 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.899583 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.899604 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:42 crc kubenswrapper[4800]: I0122 00:07:42.899622 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:42Z","lastTransitionTime":"2026-01-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.004536 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.004593 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.004609 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.004635 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.004652 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.106824 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.106868 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.106876 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.106906 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.106916 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.210828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.210918 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.210936 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.210958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.210976 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.338988 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.339044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.339062 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.339108 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.339123 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.430089 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.430137 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.430148 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.430164 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.430175 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.450864 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:43Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.455274 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.455470 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.455565 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.455656 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.455738 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.471516 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:43Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.477241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.477325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.477371 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.477389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.477402 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.492334 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:43Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.498188 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.498234 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.498249 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.498265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.498274 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.515337 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:43Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.520193 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.520249 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.520267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.520289 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.520308 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.541500 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:43Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:43 crc kubenswrapper[4800]: E0122 00:07:43.541666 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.543956 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.544000 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.544018 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.544043 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.544060 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.647004 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.647063 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.647081 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.647108 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.647124 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.750275 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.750333 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.750345 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.750366 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.750378 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.786117 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 17:01:44.522352036 +0000 UTC Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.854337 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.854400 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.854411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.854433 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.854448 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.957216 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.957302 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.957324 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.957361 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:43 crc kubenswrapper[4800]: I0122 00:07:43.957391 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:43Z","lastTransitionTime":"2026-01-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.060181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.060233 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.060249 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.060270 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.060284 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.164063 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.164135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.164153 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.164181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.164203 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.267872 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.267958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.267974 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.268005 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.268020 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.371148 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.371218 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.371239 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.371267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.371287 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.474988 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.475034 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.475046 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.475071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.475085 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.580522 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.580634 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.580659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.580695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.580720 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.684644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.684707 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.684724 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.684750 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.684768 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.786550 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 06:46:00.431185088 +0000 UTC Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.788900 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.788934 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.788944 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.788960 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.788968 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.818006 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.818212 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.818407 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:44 crc kubenswrapper[4800]: E0122 00:07:44.818403 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.818493 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:44 crc kubenswrapper[4800]: E0122 00:07:44.818561 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:44 crc kubenswrapper[4800]: E0122 00:07:44.818764 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:44 crc kubenswrapper[4800]: E0122 00:07:44.819153 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.837410 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.863660 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.884121 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.893388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.893456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.893478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.893509 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.893528 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.910294 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.938381 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.960663 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.979864 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.996573 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:44Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.998863 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.998945 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.998961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.998988 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:44 crc kubenswrapper[4800]: I0122 00:07:44.999003 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:44Z","lastTransitionTime":"2026-01-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.017304 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.030065 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.046653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.061568 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.073876 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.090501 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.102733 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.102819 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.102840 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.102923 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.102947 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.116166 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.134280 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.158641 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:45Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.206360 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.206414 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.206424 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.206439 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.206457 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.309594 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.309659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.309669 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.309685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.309697 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.360669 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:45 crc kubenswrapper[4800]: E0122 00:07:45.360945 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:45 crc kubenswrapper[4800]: E0122 00:07:45.361052 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:07:53.361030013 +0000 UTC m=+48.841333051 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.412813 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.412863 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.412873 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.412906 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.412916 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.515922 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.515995 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.516017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.516043 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.516065 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.619771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.619840 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.619854 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.619905 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.619918 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.723617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.723655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.723665 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.723682 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.723691 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.787346 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 00:52:21.415742257 +0000 UTC Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.826699 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.826943 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.826985 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.827026 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.827050 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.930020 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.930066 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.930079 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.930098 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:45 crc kubenswrapper[4800]: I0122 00:07:45.930111 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:45Z","lastTransitionTime":"2026-01-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.034344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.034390 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.034467 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.034486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.034653 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.138029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.138086 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.138101 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.138124 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.138144 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.240188 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.240227 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.240237 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.240255 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.240264 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.342316 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.342392 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.342416 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.342446 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.342468 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.445258 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.445313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.445331 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.445352 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.445366 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.548775 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.548839 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.548858 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.548923 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.548954 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.652449 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.652494 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.652504 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.652526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.652536 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.756315 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.756588 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.756615 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.756653 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.756678 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.788060 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 04:08:16.68937627 +0000 UTC Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.817673 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.817685 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.818378 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:46 crc kubenswrapper[4800]: E0122 00:07:46.818572 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.818693 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:46 crc kubenswrapper[4800]: E0122 00:07:46.818940 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:46 crc kubenswrapper[4800]: E0122 00:07:46.819179 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:46 crc kubenswrapper[4800]: E0122 00:07:46.819309 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.860379 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.860456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.860472 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.860499 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.860520 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.963570 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.963644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.963663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.963689 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:46 crc kubenswrapper[4800]: I0122 00:07:46.963706 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:46Z","lastTransitionTime":"2026-01-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.066525 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.066611 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.066631 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.066664 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.066685 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.169347 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.169391 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.169401 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.169415 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.169426 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.271574 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.271613 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.271621 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.271642 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.271652 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.375625 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.375677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.375693 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.375717 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.375733 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.479001 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.480168 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.480420 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.480677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.480851 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.584705 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.584751 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.584763 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.584780 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.584790 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.687299 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.687768 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.687910 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.688011 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.688125 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.788761 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 03:07:15.302661389 +0000 UTC Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.791102 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.791145 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.791162 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.791186 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.791204 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.819723 4800 scope.go:117] "RemoveContainer" containerID="6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.893365 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.893676 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.893689 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.893707 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.893719 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.996154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.996214 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.996232 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.996258 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:47 crc kubenswrapper[4800]: I0122 00:07:47.996277 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:47Z","lastTransitionTime":"2026-01-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.099818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.099843 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.099851 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.099864 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.099872 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.161707 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/1.log" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.165935 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.166392 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.200417 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.203276 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.203325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.203338 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.203357 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.203370 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.214448 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.230521 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.255424 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.276154 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.289033 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.301412 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.306441 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.306498 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.306513 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.306539 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.306560 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.323109 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.348160 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.373347 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.389551 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.404162 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.409605 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.409653 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.409663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.409679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.409688 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.421136 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.434415 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.448261 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.459967 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.472199 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.512632 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.512693 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.512705 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.512722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.512734 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.615303 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.615349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.615359 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.615378 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.615389 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.717514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.717570 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.717585 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.717606 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.717618 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.788943 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 19:13:52.16671107 +0000 UTC Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.817489 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.817586 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.817618 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.817504 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:48 crc kubenswrapper[4800]: E0122 00:07:48.817776 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:48 crc kubenswrapper[4800]: E0122 00:07:48.817987 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:48 crc kubenswrapper[4800]: E0122 00:07:48.818365 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:48 crc kubenswrapper[4800]: E0122 00:07:48.818488 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.820034 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.820105 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.820128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.820153 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.820175 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.923201 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.923286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.923297 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.923315 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:48 crc kubenswrapper[4800]: I0122 00:07:48.923324 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:48Z","lastTransitionTime":"2026-01-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.025663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.025720 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.025734 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.025755 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.025768 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.128879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.128948 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.128961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.128982 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.128995 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.172975 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/2.log" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.173832 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/1.log" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.178571 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" exitCode=1 Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.178616 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.178668 4800 scope.go:117] "RemoveContainer" containerID="6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.179676 4800 scope.go:117] "RemoveContainer" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" Jan 22 00:07:49 crc kubenswrapper[4800]: E0122 00:07:49.179985 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.198657 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.217072 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.232100 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.232159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.232177 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.232204 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.232222 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.240298 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.259078 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.275224 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.295636 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.321330 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6194627ae2c20a9289df27851c6488883d5d6ea9339db02a1035fbc186dff79d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:34Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI0122 00:07:34.110831 6218 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 00:07:34.110967 6218 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0122 00:07:34.111376 6218 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0122 00:07:34.112071 6218 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 00:07:34.112174 6218 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 00:07:34.112263 6218 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 00:07:34.112337 6218 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 00:07:34.112413 6218 factory.go:656] Stopping watch factory\\\\nI0122 00:07:34.112488 6218 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 00:07:34.112555 6218 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 00:07:34.112636 6218 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 00:07:34.112652 6218 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.337956 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.339485 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.339539 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.339555 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.339582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.339597 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.353538 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.368191 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.395663 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.411575 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.428225 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.442387 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.442910 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.443029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.443138 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.443229 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.447778 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.467123 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.487814 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.506462 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:49Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.547433 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.548029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.548061 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.548084 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.548099 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.651720 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.651779 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.651796 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.651818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.651835 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.754156 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.754244 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.754295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.754319 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.754333 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.789932 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 17:27:58.198940293 +0000 UTC Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.858079 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.858133 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.858165 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.858183 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.858194 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.961334 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.961412 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.961434 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.961465 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:49 crc kubenswrapper[4800]: I0122 00:07:49.961487 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:49Z","lastTransitionTime":"2026-01-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.064954 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.065090 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.065121 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.065172 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.065223 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.176311 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.176395 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.176416 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.176449 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.176470 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.185392 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/2.log" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.191119 4800 scope.go:117] "RemoveContainer" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" Jan 22 00:07:50 crc kubenswrapper[4800]: E0122 00:07:50.191722 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.214285 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.234867 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.264087 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.280215 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.280519 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.280687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.280825 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.280994 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.285227 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.304590 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.328283 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.353179 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.379331 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.384847 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.384921 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.384935 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.384954 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.384966 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.402550 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.421186 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.442277 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.461118 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.484000 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.489092 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.489143 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.489166 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.489198 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.489222 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.503653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.517129 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.548383 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.565568 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:50Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.593190 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.593245 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.593257 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.593277 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.593290 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.696241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.696296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.696309 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.696325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.696335 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.791179 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 13:31:06.0152773 +0000 UTC Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.800450 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.800530 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.800549 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.800577 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.800599 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.818404 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.818455 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.818504 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:50 crc kubenswrapper[4800]: E0122 00:07:50.818674 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:50 crc kubenswrapper[4800]: E0122 00:07:50.818843 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.818997 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:50 crc kubenswrapper[4800]: E0122 00:07:50.819040 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:50 crc kubenswrapper[4800]: E0122 00:07:50.819251 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.903507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.903577 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.903595 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.903622 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:50 crc kubenswrapper[4800]: I0122 00:07:50.903644 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:50Z","lastTransitionTime":"2026-01-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.006810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.006906 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.006924 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.006953 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.006976 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.110593 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.110638 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.110655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.110682 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.110697 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.213710 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.213802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.213818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.213837 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.213851 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.316586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.316647 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.316669 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.316700 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.316721 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.419803 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.420135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.420200 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.420259 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.420319 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.522304 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.522334 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.522341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.522353 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.522362 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.625084 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.625157 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.625175 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.625601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.625631 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.729738 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.729817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.729840 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.729872 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.730002 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.792159 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 07:47:46.371290254 +0000 UTC Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.833828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.833922 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.833944 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.833979 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.834021 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.936829 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.936919 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.936940 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.936972 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:51 crc kubenswrapper[4800]: I0122 00:07:51.936993 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:51Z","lastTransitionTime":"2026-01-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.040925 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.041014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.041036 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.041065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.041086 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.143984 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.144018 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.144029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.144047 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.144056 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.247220 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.247290 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.247308 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.247341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.247366 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.350550 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.350612 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.350623 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.350642 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.350655 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.453679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.453762 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.453780 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.453805 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.453824 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.556687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.556730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.556739 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.556754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.556763 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.659955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.660064 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.660087 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.660118 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.660136 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.763614 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.763695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.763714 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.763752 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.763772 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.793220 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 17:39:31.679055704 +0000 UTC Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.818032 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.818136 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.818167 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.818044 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:52 crc kubenswrapper[4800]: E0122 00:07:52.818323 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:52 crc kubenswrapper[4800]: E0122 00:07:52.818506 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:52 crc kubenswrapper[4800]: E0122 00:07:52.818704 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:52 crc kubenswrapper[4800]: E0122 00:07:52.818917 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.867629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.867796 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.867822 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.867882 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.867948 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.972273 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.972699 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.972879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.973106 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:52 crc kubenswrapper[4800]: I0122 00:07:52.973299 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:52Z","lastTransitionTime":"2026-01-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.077039 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.077445 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.077601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.077761 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.077928 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.181341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.181437 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.181455 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.181482 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.181500 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.284995 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.285066 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.285085 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.285115 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.285136 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.388099 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.388212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.388238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.388279 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.388307 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.454685 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.454772 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.455067 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:09.455040796 +0000 UTC m=+64.935343834 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.492023 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.492130 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.492545 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.493153 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.493213 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.597161 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.597257 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.597284 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.597322 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.597352 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.688120 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.688236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.688263 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.688302 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.688332 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.708437 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:53Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.714290 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.714366 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.714385 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.714411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.714433 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.729750 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:53Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.735088 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.735165 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.735185 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.735208 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.735222 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.749973 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:53Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.754996 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.755040 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.755057 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.755074 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.755086 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.772123 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:53Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.776498 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.776532 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.776541 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.776553 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.776564 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.789792 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:53Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:53 crc kubenswrapper[4800]: E0122 00:07:53.790059 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.792455 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.792518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.792533 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.792553 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.792566 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.793427 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 12:35:47.278938005 +0000 UTC Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.895544 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.895628 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.895672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.895695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.895708 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.999398 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.999497 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.999522 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.999562 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:53 crc kubenswrapper[4800]: I0122 00:07:53.999592 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:53Z","lastTransitionTime":"2026-01-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.103340 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.103433 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.103452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.103478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.103496 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.206090 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.206173 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.206200 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.206238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.206261 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.309608 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.309677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.309695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.309723 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.309741 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.413619 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.413677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.413696 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.413721 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.413740 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.516674 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.516740 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.516757 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.516786 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.516808 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.620539 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.620641 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.620681 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.620720 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.620746 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.724073 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.724132 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.724146 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.724170 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.724187 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.774172 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.774387 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774424 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:08:26.774373557 +0000 UTC m=+82.254676635 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.774546 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.774612 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774621 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.774675 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774766 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:26.774733276 +0000 UTC m=+82.255036344 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774813 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774817 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774847 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774870 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774815 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.774967 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:26.774944972 +0000 UTC m=+82.255248050 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.775006 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:26.774989153 +0000 UTC m=+82.255292481 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.775039 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.775080 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.775197 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:26.775161798 +0000 UTC m=+82.255465036 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.794488 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 00:59:13.899650758 +0000 UTC Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.818282 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.818332 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.818495 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.818580 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.818834 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.818838 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.819025 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:54 crc kubenswrapper[4800]: E0122 00:07:54.819194 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.827494 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.827564 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.827588 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.827617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.827639 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.837620 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.857845 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.912924 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.942657 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.947389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.947466 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.947493 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.947531 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.947559 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:54Z","lastTransitionTime":"2026-01-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:54 crc kubenswrapper[4800]: I0122 00:07:54.979218 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.005336 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:54Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.020496 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.034030 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.047689 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.051458 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.051509 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.051526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.051552 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.051568 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.060865 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.095951 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.110497 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.124602 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.138206 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.154932 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.154996 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.155012 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.155036 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.155088 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.155094 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.171677 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.186405 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:55Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.259921 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.259973 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.259984 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.260001 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.260012 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.363242 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.363305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.363324 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.363347 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.363363 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.466687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.466797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.466816 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.466850 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.466874 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.570533 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.570610 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.570631 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.570659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.570682 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.674110 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.674174 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.674187 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.674212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.674227 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.777764 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.777851 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.777878 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.777957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.777985 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.794604 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:25:00.659961523 +0000 UTC Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.881735 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.881817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.881844 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.881915 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.881946 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.986195 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.986282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.986306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.986350 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:55 crc kubenswrapper[4800]: I0122 00:07:55.986376 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:55Z","lastTransitionTime":"2026-01-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.089496 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.089576 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.089594 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.089625 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.089664 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.193192 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.193242 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.193259 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.193286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.193304 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.297758 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.297848 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.297874 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.297951 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.297979 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.402614 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.402698 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.402718 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.402754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.402778 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.506865 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.507207 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.507293 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.507327 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.507349 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.610675 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.610743 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.610761 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.610790 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.610815 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.714947 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.715006 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.715017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.715036 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.715049 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.795233 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:26:12.570103434 +0000 UTC Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.817459 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.817482 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:56 crc kubenswrapper[4800]: E0122 00:07:56.817597 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.817609 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.817673 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:56 crc kubenswrapper[4800]: E0122 00:07:56.817771 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:56 crc kubenswrapper[4800]: E0122 00:07:56.817942 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:56 crc kubenswrapper[4800]: E0122 00:07:56.818171 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.818517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.818591 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.818613 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.818647 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.818667 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.922706 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.922786 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.922810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.922850 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.922879 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:56Z","lastTransitionTime":"2026-01-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.932248 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.949800 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:56Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.951487 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.969572 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:56Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:56 crc kubenswrapper[4800]: I0122 00:07:56.983628 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:56Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.012712 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.026427 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.026538 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.026559 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.026629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.026654 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.034177 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.058396 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.078519 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.104275 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.120222 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.129807 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.129847 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.129859 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.129875 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.129906 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.137765 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.155833 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.176693 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.194878 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.211234 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.223513 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.232430 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.232476 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.232498 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.232524 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.232543 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.246259 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.273372 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:57Z is after 2025-08-24T17:21:41Z" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.336997 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.337076 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.337091 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.337113 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.337129 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.440299 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.440373 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.440393 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.440423 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.440446 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.543492 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.543583 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.543616 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.543655 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.543686 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.646422 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.646486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.646504 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.646524 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.646536 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.750050 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.750101 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.750112 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.750131 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.750144 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.795671 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 14:15:29.114121801 +0000 UTC Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.853033 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.853104 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.853129 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.853158 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.853184 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.956170 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.956246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.956272 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.956305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:57 crc kubenswrapper[4800]: I0122 00:07:57.956329 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:57Z","lastTransitionTime":"2026-01-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.059852 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.059951 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.059976 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.060014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.060040 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.163528 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.163580 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.163595 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.163619 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.163636 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.267349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.267425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.267438 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.267464 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.267479 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.371983 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.372426 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.372663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.372856 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.373083 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.477319 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.477731 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.478316 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.478732 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.479178 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.582526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.582934 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.583161 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.583366 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.583599 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.686578 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.686617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.686627 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.686644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.686655 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.790808 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.791358 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.791563 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.791754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.791954 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.796199 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 19:24:13.023311104 +0000 UTC Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.818115 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:07:58 crc kubenswrapper[4800]: E0122 00:07:58.818377 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.818544 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:07:58 crc kubenswrapper[4800]: E0122 00:07:58.818832 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.818696 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:07:58 crc kubenswrapper[4800]: E0122 00:07:58.819136 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.819286 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:07:58 crc kubenswrapper[4800]: E0122 00:07:58.819429 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.895454 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.895903 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.896157 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.896296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:58 crc kubenswrapper[4800]: I0122 00:07:58.896407 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:58Z","lastTransitionTime":"2026-01-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.000309 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.000702 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.000814 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.000958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.001059 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.104947 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.104993 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.105008 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.105034 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.105048 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.208660 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.208759 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.208788 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.208825 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.208851 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.313100 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.313187 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.313207 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.313238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.313258 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.415746 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.415820 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.415843 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.415873 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.415928 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.518241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.518299 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.518315 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.518341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.518361 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.621144 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.621199 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.621214 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.621236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.621252 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.723787 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.723838 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.723852 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.723873 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.723905 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.796495 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 10:22:44.543566246 +0000 UTC Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.827507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.827554 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.827570 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.827588 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.827603 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.930738 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.930806 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.930818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.930845 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:07:59 crc kubenswrapper[4800]: I0122 00:07:59.930857 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:07:59Z","lastTransitionTime":"2026-01-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.033967 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.034062 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.034086 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.034115 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.034134 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.137727 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.137801 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.137818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.137848 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.137867 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.240111 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.240213 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.240232 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.240706 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.241072 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.345149 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.345243 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.345271 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.345310 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.345338 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.448771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.448842 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.448879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.448950 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.448977 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.552437 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.552481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.552489 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.552503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.552514 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.655842 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.655944 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.655962 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.655991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.656010 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.759922 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.759996 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.760012 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.760039 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.760058 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.797788 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 16:27:12.065132622 +0000 UTC Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.817641 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.817741 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:00 crc kubenswrapper[4800]: E0122 00:08:00.817872 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.817970 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:00 crc kubenswrapper[4800]: E0122 00:08:00.818057 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:00 crc kubenswrapper[4800]: E0122 00:08:00.818218 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.818715 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:00 crc kubenswrapper[4800]: E0122 00:08:00.818883 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.864143 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.864214 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.864236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.864269 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.864289 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.967730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.967793 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.967809 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.967835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:00 crc kubenswrapper[4800]: I0122 00:08:00.967851 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:00Z","lastTransitionTime":"2026-01-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.071097 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.071568 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.071820 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.072101 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.072331 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.176858 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.176941 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.176958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.176981 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.176996 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.280575 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.280629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.280643 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.280666 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.280678 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.383522 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.383594 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.383608 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.383629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.383658 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.487730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.487823 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.487913 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.487955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.487982 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.592726 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.592814 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.592835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.592870 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.592929 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.697042 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.697593 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.697776 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.698025 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.698270 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.798233 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:37:32.539274198 +0000 UTC Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.802675 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.802729 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.802746 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.802775 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.802794 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.906178 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.906241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.906257 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.906284 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:01 crc kubenswrapper[4800]: I0122 00:08:01.906303 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:01Z","lastTransitionTime":"2026-01-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.010150 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.010351 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.010391 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.010422 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.010446 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.113660 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.113749 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.113769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.113796 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.113817 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.217715 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.217799 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.217821 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.217854 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.217876 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.321344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.321421 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.321440 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.321473 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.321492 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.424301 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.424421 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.424440 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.424471 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.424490 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.527536 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.527624 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.527654 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.527694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.527726 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.630617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.630686 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.630701 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.630727 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.630741 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.733124 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.733180 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.733194 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.733215 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.733232 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.799545 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 14:50:06.546759219 +0000 UTC Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.818029 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.818103 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:02 crc kubenswrapper[4800]: E0122 00:08:02.818258 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.818385 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.818444 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:02 crc kubenswrapper[4800]: E0122 00:08:02.818583 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:02 crc kubenswrapper[4800]: E0122 00:08:02.818473 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:02 crc kubenswrapper[4800]: E0122 00:08:02.818805 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.835716 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.835786 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.835804 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.835830 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.835850 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.939853 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.939939 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.939955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.939975 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:02 crc kubenswrapper[4800]: I0122 00:08:02.939988 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:02Z","lastTransitionTime":"2026-01-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.042681 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.042721 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.042730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.042745 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.042756 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.145368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.145430 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.145444 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.145472 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.145488 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.248695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.248742 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.248756 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.248772 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.248781 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.352088 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.352142 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.352152 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.352172 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.352183 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.455123 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.455205 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.455241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.455272 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.455295 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.558682 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.558748 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.558760 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.558810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.558828 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.661879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.662000 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.662020 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.662053 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.662072 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.765478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.765554 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.765606 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.765646 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.765672 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.800516 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 03:59:17.104388347 +0000 UTC Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.819039 4800 scope.go:117] "RemoveContainer" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" Jan 22 00:08:03 crc kubenswrapper[4800]: E0122 00:08:03.819292 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.869259 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.869304 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.869321 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.869345 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.869359 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.907055 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.907140 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.907164 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.907197 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.907218 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: E0122 00:08:03.928831 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:03Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.933731 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.933762 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.933802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.933828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.933845 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: E0122 00:08:03.950263 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:03Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.954503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.954563 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.954582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.954607 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.954627 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:03 crc kubenswrapper[4800]: E0122 00:08:03.975459 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:03Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.981508 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.981579 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.981597 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.981624 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:03 crc kubenswrapper[4800]: I0122 00:08:03.981643 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:03Z","lastTransitionTime":"2026-01-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.002359 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:03Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.007596 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.007659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.007674 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.007728 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.007748 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.028243 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.028416 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.030337 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.030370 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.030383 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.030419 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.030431 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.134169 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.134244 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.134258 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.134278 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.134293 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.238452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.238542 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.238564 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.238603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.238638 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.341742 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.341826 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.341848 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.341877 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.341934 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.444403 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.444468 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.444479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.444518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.444529 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.548403 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.548479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.548502 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.548531 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.548556 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.652866 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.652988 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.653014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.653045 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.653063 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.755785 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.755923 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.755944 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.755977 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.755997 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.801187 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 01:07:23.938753605 +0000 UTC Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.817627 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.817703 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.817715 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.817832 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.818006 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.818110 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.818188 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:04 crc kubenswrapper[4800]: E0122 00:08:04.818247 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.840446 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.860082 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.860156 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.860174 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.860202 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.860221 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.861082 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.882145 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.902091 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.918375 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.933304 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.945070 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.961619 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.963263 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.963325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.963338 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.963356 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.963366 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:04Z","lastTransitionTime":"2026-01-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.977553 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:04 crc kubenswrapper[4800]: I0122 00:08:04.994724 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:04Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.012606 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.025474 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.040147 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066770 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066821 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066842 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066855 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.066740 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.080482 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.098619 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.111444 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.130328 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:05Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.170720 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.171142 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.171333 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.171526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.171705 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.275245 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.275324 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.275342 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.275367 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.275381 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.378007 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.378038 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.378046 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.378082 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.378091 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.481342 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.481399 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.481411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.481434 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.481447 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.590425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.590489 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.590506 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.590534 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.590553 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.694661 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.694728 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.694738 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.694751 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.694782 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.797300 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.797344 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.797353 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.797369 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.797381 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.801725 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 22:18:10.905943413 +0000 UTC Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.900569 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.900672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.900692 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.900719 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:05 crc kubenswrapper[4800]: I0122 00:08:05.900739 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:05Z","lastTransitionTime":"2026-01-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.003578 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.003613 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.003623 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.003636 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.003645 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.106184 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.106261 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.106278 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.106313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.106334 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.209051 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.209124 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.209144 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.209177 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.209202 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.312568 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.312751 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.312783 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.312823 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.312853 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.415507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.415584 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.415603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.415634 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.415654 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.518670 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.518766 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.518783 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.518813 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.518834 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.622045 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.622123 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.622141 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.622171 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.622202 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.725440 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.725503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.725521 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.725557 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.725575 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.802833 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 06:20:38.785042817 +0000 UTC Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.817510 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.817602 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.817661 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:06 crc kubenswrapper[4800]: E0122 00:08:06.817680 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.817537 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:06 crc kubenswrapper[4800]: E0122 00:08:06.817832 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:06 crc kubenswrapper[4800]: E0122 00:08:06.817871 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:06 crc kubenswrapper[4800]: E0122 00:08:06.817965 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.828478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.828518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.828527 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.828544 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.828556 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.930564 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.930600 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.930608 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.930623 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:06 crc kubenswrapper[4800]: I0122 00:08:06.930632 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:06Z","lastTransitionTime":"2026-01-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.033216 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.033270 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.033282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.033301 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.033314 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.136358 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.136522 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.136545 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.136619 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.136657 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.239060 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.239143 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.239157 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.239178 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.239198 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.342479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.342563 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.342587 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.342613 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.342631 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.445201 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.445251 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.445267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.445289 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.445303 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.548298 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.548365 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.548374 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.548395 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.548413 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.651044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.651103 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.651112 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.651128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.651138 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.755685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.755749 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.755770 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.755797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.755815 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.803504 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 05:24:13.882655431 +0000 UTC Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.858527 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.858571 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.858582 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.858600 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.858611 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.961715 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.961769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.961778 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.961795 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:07 crc kubenswrapper[4800]: I0122 00:08:07.961806 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:07Z","lastTransitionTime":"2026-01-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.064142 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.064184 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.064195 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.064210 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.064220 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.167147 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.167186 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.167196 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.167212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.167221 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.269731 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.269774 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.269785 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.269802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.269814 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.373365 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.373454 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.373477 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.373508 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.373527 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.477178 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.477248 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.477262 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.477282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.477295 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.579947 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.579987 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.579998 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.580014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.580026 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.683460 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.683508 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.683520 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.683537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.683551 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.812918 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 14:00:16.189105352 +0000 UTC Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.815025 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.815150 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.815173 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.815211 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.815234 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.817477 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.817525 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.817590 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.817487 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:08 crc kubenswrapper[4800]: E0122 00:08:08.817701 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:08 crc kubenswrapper[4800]: E0122 00:08:08.817907 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:08 crc kubenswrapper[4800]: E0122 00:08:08.818030 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:08 crc kubenswrapper[4800]: E0122 00:08:08.818106 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.918385 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.918654 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.918723 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.918811 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:08 crc kubenswrapper[4800]: I0122 00:08:08.918921 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:08Z","lastTransitionTime":"2026-01-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.022383 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.022434 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.022445 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.022461 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.022472 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.125267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.125334 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.125346 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.125364 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.125375 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.228645 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.228694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.228707 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.228726 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.228736 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.332142 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.332207 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.332223 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.332242 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.332256 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.435577 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.435934 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.436055 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.436131 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.436221 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.539305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.539346 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.539359 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.539381 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.539394 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.548441 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:09 crc kubenswrapper[4800]: E0122 00:08:09.548693 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:08:09 crc kubenswrapper[4800]: E0122 00:08:09.548798 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:08:41.548770274 +0000 UTC m=+97.029073322 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.642668 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.642716 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.642726 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.642743 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.642761 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.745664 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.745713 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.745724 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.745741 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.745753 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.813975 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 15:32:47.279770159 +0000 UTC Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.848181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.848601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.848708 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.848817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.848940 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.951939 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.951988 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.951999 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.952023 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:09 crc kubenswrapper[4800]: I0122 00:08:09.952036 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:09Z","lastTransitionTime":"2026-01-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.055601 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.055644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.055654 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.055678 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.055690 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.158954 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.159264 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.159323 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.159401 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.159459 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.262694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.263288 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.263478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.263608 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.263733 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.269535 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/0.log" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.269613 4800 generic.go:334] "Generic (PLEG): container finished" podID="aff2b22c-b87d-47c9-b5a0-6e5772022488" containerID="7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a" exitCode=1 Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.269666 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerDied","Data":"7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.270384 4800 scope.go:117] "RemoveContainer" containerID="7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.292736 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.323075 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.339200 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.354924 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.368632 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.369621 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.369759 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.369871 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.369993 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.370075 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.392449 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.408202 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.425861 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.442159 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.457783 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.471647 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.473780 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.473836 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.473851 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.473877 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.473913 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.487787 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.501916 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.515588 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.530161 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.548677 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.560983 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.574051 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:10Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.576110 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.576197 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.576225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.576265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.576300 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.679311 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.679387 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.679407 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.679438 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.679458 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.782464 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.782518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.782533 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.782554 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.782567 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.815875 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 14:15:13.955236843 +0000 UTC Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.817306 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.817388 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.817440 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.817438 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:10 crc kubenswrapper[4800]: E0122 00:08:10.817515 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:10 crc kubenswrapper[4800]: E0122 00:08:10.817742 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:10 crc kubenswrapper[4800]: E0122 00:08:10.818003 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:10 crc kubenswrapper[4800]: E0122 00:08:10.818189 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.885839 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.886379 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.886528 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.886658 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.886761 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.989685 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.989986 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.990052 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.990116 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:10 crc kubenswrapper[4800]: I0122 00:08:10.990211 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:10Z","lastTransitionTime":"2026-01-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.093385 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.093465 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.093483 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.093510 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.093529 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.196282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.196366 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.196388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.196422 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.196445 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.275439 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/0.log" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.275512 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerStarted","Data":"7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.292604 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.299141 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.299188 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.299201 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.299222 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.299232 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.309653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.324157 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.348732 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.366159 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.381564 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.400717 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.402244 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.402404 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.402532 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.402651 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.402798 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.415494 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.434600 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.453698 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.466711 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.483620 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.496586 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.506001 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.506059 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.506073 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.506093 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.506106 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.512785 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.526711 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.541191 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.551605 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.569229 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:11Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.610865 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.610967 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.610992 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.611026 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.611049 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.716855 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.717019 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.717040 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.717074 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.717090 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.817134 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 09:09:45.199311976 +0000 UTC Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.820059 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.820129 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.820151 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.820182 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.820200 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.923203 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.923255 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.923266 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.923282 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:11 crc kubenswrapper[4800]: I0122 00:08:11.923292 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:11Z","lastTransitionTime":"2026-01-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.026196 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.026241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.026254 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.026276 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.026290 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.129073 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.129115 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.129125 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.129140 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.129152 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.232556 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.232695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.232717 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.232747 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.232768 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.335796 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.335869 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.335919 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.335956 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.335981 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.439162 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.439226 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.439244 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.439271 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.439290 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.542232 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.542294 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.542312 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.542340 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.542358 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.645140 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.645212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.645225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.645244 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.645255 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.748269 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.748317 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.748329 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.748349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.748363 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.817347 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 01:38:28.407845686 +0000 UTC Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.817624 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.817630 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:12 crc kubenswrapper[4800]: E0122 00:08:12.818086 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.817642 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.817630 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:12 crc kubenswrapper[4800]: E0122 00:08:12.818386 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:12 crc kubenswrapper[4800]: E0122 00:08:12.818467 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:12 crc kubenswrapper[4800]: E0122 00:08:12.818528 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.850673 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.850714 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.850724 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.850742 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.850757 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.953993 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.954033 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.954043 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.954060 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:12 crc kubenswrapper[4800]: I0122 00:08:12.954070 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:12Z","lastTransitionTime":"2026-01-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.062429 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.062562 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.062589 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.062616 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.062633 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.165961 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.166014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.166028 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.166052 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.166068 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.269561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.269628 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.269646 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.269673 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.269693 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.372823 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.372855 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.372863 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.372874 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.372902 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.476349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.476411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.476423 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.476442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.476455 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.579368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.579439 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.579452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.579477 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.579494 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.682429 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.682518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.682530 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.682548 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.682558 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.790552 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.790903 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.790919 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.790955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.790968 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.818448 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 14:26:20.514213398 +0000 UTC Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.894119 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.894165 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.894175 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.894194 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.894208 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.997568 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.997633 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.997650 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.997678 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:13 crc kubenswrapper[4800]: I0122 00:08:13.997694 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:13Z","lastTransitionTime":"2026-01-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.100959 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.101032 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.101053 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.101083 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.101101 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.204313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.204639 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.204765 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.204861 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.204970 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.308199 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.308687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.308920 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.309702 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.309940 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.329832 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.329911 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.329923 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.329943 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.329960 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.348926 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.353479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.353518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.353532 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.353553 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.353567 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.372550 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.377702 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.377739 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.377750 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.377771 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.377786 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.395961 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.400287 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.400330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.400342 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.400364 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.400378 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.417091 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.422732 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.422810 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.422835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.422870 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.422929 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.453539 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.453775 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.458536 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.458581 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.458594 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.458617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.458632 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.561135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.561181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.561193 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.561213 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.561223 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.663925 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.663965 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.663976 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.663996 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.664012 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.766992 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.767767 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.767862 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.767957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.768034 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.817194 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.817246 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.817211 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.817353 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.817419 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.817529 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.817613 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:14 crc kubenswrapper[4800]: E0122 00:08:14.817671 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.819151 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 02:30:52.629172677 +0000 UTC Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.838464 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.855406 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.868158 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.870644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.870682 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.870695 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.870717 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.870731 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.889403 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.905099 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.924201 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.941675 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.954477 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.971704 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.973558 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.973591 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.973603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.973626 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.973641 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:14Z","lastTransitionTime":"2026-01-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.982608 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:14 crc kubenswrapper[4800]: I0122 00:08:14.992954 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:14Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.005337 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.024535 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.037005 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.049619 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.076737 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.076764 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.076772 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.076791 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.076800 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.179971 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.180012 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.180025 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.180045 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.180060 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.283178 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.283215 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.283229 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.283250 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.283264 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.387415 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.387468 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.387486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.387513 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.387532 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.405842 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.427293 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.439360 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:15Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.490332 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.490400 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.490418 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.490448 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.490467 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.593870 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.593930 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.593939 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.593957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.593968 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.697135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.697197 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.697215 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.697246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.697265 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.800388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.800431 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.800441 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.800457 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.800468 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.818550 4800 scope.go:117] "RemoveContainer" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.819354 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 23:09:52.079254074 +0000 UTC Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.903238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.903286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.903295 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.903314 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:15 crc kubenswrapper[4800]: I0122 00:08:15.903327 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:15Z","lastTransitionTime":"2026-01-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.006151 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.006642 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.006663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.006694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.006714 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.110046 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.110102 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.110114 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.110132 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.110142 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.212409 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.212447 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.212456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.212470 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.212480 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.314435 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.314485 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.314496 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.314511 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.314522 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.412085 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/2.log" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.414625 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.416510 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.420217 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.420283 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.420306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.420335 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.420357 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.433237 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.447829 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.463092 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.475936 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.486266 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.499632 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.521426 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.523225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.523259 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.523273 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.523297 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.523312 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.536063 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.552469 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.565076 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.604433 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.618533 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.625615 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.625657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.625672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.625693 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.625706 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.634798 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.649548 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.662459 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.679004 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.693209 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.704971 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:16Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.728555 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.728592 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.728603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.728618 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.728627 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.817346 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.817482 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.817347 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.817565 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:16 crc kubenswrapper[4800]: E0122 00:08:16.817492 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:16 crc kubenswrapper[4800]: E0122 00:08:16.817703 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:16 crc kubenswrapper[4800]: E0122 00:08:16.817823 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:16 crc kubenswrapper[4800]: E0122 00:08:16.817925 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.820369 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 10:16:30.735923708 +0000 UTC Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.831398 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.831452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.831471 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.831497 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.831520 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.935375 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.935433 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.935450 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.935477 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:16 crc kubenswrapper[4800]: I0122 00:08:16.935498 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:16Z","lastTransitionTime":"2026-01-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.038425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.038490 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.038507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.038534 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.038554 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.141920 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.141975 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.141989 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.142013 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.142028 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.245241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.245296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.245313 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.245337 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.245349 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.348148 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.348205 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.348217 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.348241 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.348256 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.420910 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/3.log" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.421635 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/2.log" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.424954 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" exitCode=1 Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.425005 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.425095 4800 scope.go:117] "RemoveContainer" containerID="d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.426793 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:08:17 crc kubenswrapper[4800]: E0122 00:08:17.427235 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.452802 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.452845 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.452857 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.452878 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.452907 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.453651 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.466371 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.480123 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.495584 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.517382 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.535946 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.551386 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.555632 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.555677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.555689 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.555708 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.555721 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.567191 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.588806 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.606799 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.624672 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.638471 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.654747 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.659233 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.659305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.659321 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.659345 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.659360 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.678831 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d38657bcf050b13c92fc15ded34c3cb668ecb55844d973acb76a1d69a2ed62ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:07:48Z\\\",\\\"message\\\":\\\"oSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.40:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2a3fb1a3-a476-4e14-bcf5-fb79af60206a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 00:07:48.762834 6428 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 00:07:48.763022 6428 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:07:48Z is after 2025-08-24T17:21:41Z]\\\\nI0122 00:07:48.760232 6428 services_controller.go:451] Built service openshift-operator-lifecycle-manager/packageserver-service cluster-wide LB for network=def\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:17Z\\\",\\\"message\\\":\\\"ocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 00:08:16.875061 6809 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nI0122 00:08:16.874822 6809 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI0122 00:08:16.874838 6809 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fp7s5\\\\nI0122 00:08:16.875062 6809 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0122 00:08:16.874968 6809 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0122 00:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.693711 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.706554 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.720298 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.736413 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:17Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.763180 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.763234 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.763246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.763265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.763278 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.821091 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 14:41:11.576163633 +0000 UTC Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.865518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.865575 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.865591 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.865605 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.865615 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.968832 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.968872 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.968907 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.968930 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:17 crc kubenswrapper[4800]: I0122 00:08:17.968939 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:17Z","lastTransitionTime":"2026-01-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.072162 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.072222 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.072238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.072265 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.072284 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.175525 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.175598 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.175616 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.175643 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.175663 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.279461 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.279507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.279517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.279534 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.279544 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.383046 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.383173 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.383199 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.383302 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.383402 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.431564 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/3.log" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.437179 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:08:18 crc kubenswrapper[4800]: E0122 00:08:18.437362 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.470220 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487298 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487347 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487363 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487386 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487401 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.487704 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.507044 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.525629 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.541951 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.560313 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.587653 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.590391 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.590468 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.590478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.590501 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.590516 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.600360 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.620903 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:17Z\\\",\\\"message\\\":\\\"ocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 00:08:16.875061 6809 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nI0122 00:08:16.874822 6809 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI0122 00:08:16.874838 6809 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fp7s5\\\\nI0122 00:08:16.875062 6809 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0122 00:08:16.874968 6809 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0122 00:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:08:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.637344 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.652054 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.663657 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.679003 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.692854 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.693377 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.693430 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.693448 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.693475 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.693495 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.709811 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.724076 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.737587 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.757731 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:18Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.796299 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.796356 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.796370 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.796389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.796406 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.817328 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.817453 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:18 crc kubenswrapper[4800]: E0122 00:08:18.817551 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.817661 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.817972 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:18 crc kubenswrapper[4800]: E0122 00:08:18.817953 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:18 crc kubenswrapper[4800]: E0122 00:08:18.818373 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:18 crc kubenswrapper[4800]: E0122 00:08:18.819490 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.821569 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 20:40:49.496535871 +0000 UTC Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.829503 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.899965 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.900059 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.900076 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.900119 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:18 crc kubenswrapper[4800]: I0122 00:08:18.900135 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:18Z","lastTransitionTime":"2026-01-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.004456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.004518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.004540 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.004574 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.004600 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.108503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.108588 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.108603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.108627 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.108642 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.212291 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.212395 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.212420 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.212458 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.212485 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.317015 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.317114 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.317139 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.317171 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.317195 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.420786 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.420843 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.420854 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.420874 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.420907 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.524514 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.524586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.524613 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.524645 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.524671 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.628413 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.628687 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.628713 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.628742 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.628763 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.731830 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.731938 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.731958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.731986 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.732007 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.822360 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 20:23:37.70904789 +0000 UTC Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.835730 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.835780 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.835791 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.835809 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.835820 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.938838 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.938958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.938980 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.939018 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:19 crc kubenswrapper[4800]: I0122 00:08:19.939039 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:19Z","lastTransitionTime":"2026-01-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.042118 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.042235 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.042255 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.042280 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.042298 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.145836 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.145941 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.145956 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.145978 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.145990 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.249004 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.249050 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.249065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.249082 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.249096 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.351549 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.351610 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.351628 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.351657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.351676 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.456049 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.456103 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.456121 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.456148 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.456176 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.560193 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.560307 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.560330 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.560357 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.560375 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.663603 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.663670 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.663686 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.663754 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.663777 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.767674 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.767746 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.767760 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.767792 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.767809 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.817561 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.817624 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.817696 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:20 crc kubenswrapper[4800]: E0122 00:08:20.817789 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.817816 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:20 crc kubenswrapper[4800]: E0122 00:08:20.818031 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:20 crc kubenswrapper[4800]: E0122 00:08:20.818165 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:20 crc kubenswrapper[4800]: E0122 00:08:20.818280 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.822657 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 19:40:23.34855547 +0000 UTC Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.871828 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.871871 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.871900 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.871919 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.871929 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.974994 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.975059 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.975074 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.975106 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:20 crc kubenswrapper[4800]: I0122 00:08:20.975122 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:20Z","lastTransitionTime":"2026-01-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.079479 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.079542 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.079561 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.079593 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.079608 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.183879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.183942 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.183951 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.183985 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.183999 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.287221 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.287275 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.287288 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.287307 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.287319 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.390014 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.390061 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.390072 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.390091 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.390104 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.492761 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.492827 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.492841 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.492866 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.492901 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.595916 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.595973 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.595989 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.596017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.596035 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.699580 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.699662 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.699688 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.699724 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.699747 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.803239 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.803318 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.803331 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.803355 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.803371 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.823656 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 02:21:25.468891819 +0000 UTC Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.906583 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.906629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.906638 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.906658 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:21 crc kubenswrapper[4800]: I0122 00:08:21.906669 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:21Z","lastTransitionTime":"2026-01-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.010435 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.010507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.010529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.010565 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.010589 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.114591 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.114659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.114684 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.114722 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.114744 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.218314 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.218381 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.218403 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.218432 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.218451 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.321257 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.321321 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.321341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.321369 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.321391 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.426315 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.426383 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.426401 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.426425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.426442 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.530562 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.530640 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.530658 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.530692 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.530715 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.634527 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.634605 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.634622 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.634648 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.634665 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.738360 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.738425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.738444 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.738503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.738523 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.818185 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.818257 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.818267 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:22 crc kubenswrapper[4800]: E0122 00:08:22.818448 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.818538 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:22 crc kubenswrapper[4800]: E0122 00:08:22.818742 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:22 crc kubenswrapper[4800]: E0122 00:08:22.818991 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:22 crc kubenswrapper[4800]: E0122 00:08:22.819152 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.824001 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 20:18:00.7129129 +0000 UTC Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.842371 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.842429 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.842446 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.842473 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.842494 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.946362 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.946427 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.946443 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.946476 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:22 crc kubenswrapper[4800]: I0122 00:08:22.946497 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:22Z","lastTransitionTime":"2026-01-22T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.049957 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.050060 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.050082 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.050111 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.050132 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.154708 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.154777 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.154795 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.154822 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.154841 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.263075 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.263159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.263176 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.263207 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.263226 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.367818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.367918 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.367935 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.367958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.367985 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.470799 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.470865 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.470879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.470926 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.470944 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.573308 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.573368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.573385 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.573408 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.573423 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.677035 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.677106 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.677124 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.677154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.677173 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.780235 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.780285 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.780294 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.780333 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.780342 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.824484 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 05:01:12.069580022 +0000 UTC Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.884210 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.884252 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.884261 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.884273 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.884282 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.987388 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.987486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.987517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.987555 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:23 crc kubenswrapper[4800]: I0122 00:08:23.987579 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:23Z","lastTransitionTime":"2026-01-22T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.091362 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.091453 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.091478 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.091518 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.091543 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.198054 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.198138 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.198159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.198189 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.198211 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.301615 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.301706 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.301729 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.301759 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.301783 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.404804 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.404872 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.404942 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.404995 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.405028 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.507798 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.507922 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.507950 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.507984 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.508009 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.611248 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.611296 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.611306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.611326 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.611339 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.670053 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.670128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.670157 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.670193 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.670217 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.690387 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.695616 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.695657 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.695667 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.695683 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.695694 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.712283 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.719003 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.719074 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.719093 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.719124 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.719144 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.737696 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.742402 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.742477 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.742503 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.742538 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.742563 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.763941 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.768615 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.768659 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.768670 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.768689 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.768703 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.786235 4800 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"929f49ac-7e30-48d3-8f26-0c30c68b0fdc\\\",\\\"systemUUID\\\":\\\"cbfcc22a-4122-46f9-95ef-0dbfb54ccc6f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.786400 4800 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.788411 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.788497 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.788523 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.788555 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.788580 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.817626 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.817657 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.817722 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.817854 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.817976 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.818141 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.818326 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:24 crc kubenswrapper[4800]: E0122 00:08:24.818426 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.824733 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:49:59.821176503 +0000 UTC Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.839427 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e2be9f-c2b6-4bdf-ac57-eb0add51532b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de0997a405415d6697fc0370a0fae0fcffb10933721cbc25f8735145b6cba883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef8507bdda6e6a08ca846c7b5386b1784025653fd5d09725c441a4a1acb1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://354673c6791e1e1da3fa7abc15639ec2fa594745b7c6f72ef3a31f5c172c8dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ea2ab7c18419a8325f805f3ccdc573fae1691f93749478cc3ffa33a5c84c837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.857835 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.873924 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11c615f4-8d2e-4df4-8b32-ddc8db848915\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b954f5a0981a1434fee0dc9162ca3d565f303c71c42a4e5c57fdd8f6823aec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbf8ccaf5f4e8dce2e62baaf790a859bdfa4a86f2eb510a85ba0f824492273e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fbf8ccaf5f4e8dce2e62baaf790a859bdfa4a86f2eb510a85ba0f824492273e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.891397 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.891469 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.891486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.891516 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.891535 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.906411 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3bb1d46-5d5c-43e4-b1c2-61370f69044c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38083eb85e682746fdd4142e1a8b170744fd52fb712e6e6e4af8177ea47f98b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52a0364f0273119521090904e9679307e3a446da8bb7b911c444fde2878dca4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42419665a1139704dd6fefe039128cd3f780d4d8c95b0efd84baa3f1536637d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b62080e3f42ff4eedb3f04971b229e1301509d73763a36a74fb4ad0991203c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c70f2a666f041e0c82cd5d229beb08a79009820cbe237bdcb536a6eaece529c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c130b2e354d0675f26906860cfdb5af5365b9f43f461236b598529ea65b3893\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e94c820c033798c736150d4c21a232730e73e41da300a9019eee9cbdd18964b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c1486e10780b05d0b86a6453889df029b752dce77ae80d5ed66976dc656c7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.926186 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdb45e62-a601-42e3-988e-4834834ecb6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c9946717d6053dca7407bb807f2e7087cabc003569bd196a88bc38cc4590f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98ca2274f7c0e0fe46996bbe3aa2d00732296a881e1010fc5c9f4f15ad21a72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd514c2bb01fc8dafd75c175fa5cc3f7221f534129c8079a239649b864c5dd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e97ee8a45f8fde125f6102beb94b82a3774c12bec2047f8c93a5f4a102574dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9103401a718e5db2afa163bef7f26a88f298cb6f3bb4855b1274ec7b6558b6e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://737ee9358afaa5c663719a4b1aae2ae9e1e7a8eaf2f5fda69e5dfd68fc17bb22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9676f2def622696544f301760572016258a4653ac3da4bbd8d7dbf60d6e89786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qh4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bv9wt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.945955 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7367527afb53baef514b10815ee1ca7f5da3f09156cab5b44d7bedba00ee0f7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4v8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-mrfxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.964089 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7tj9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e29e23b-f99c-4c89-8126-9493b6421e99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d01a85b86048fd60d51cb9255abfd82397d6c92520b8f44c419e85cbd6487df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rc6ll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7tj9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.981794 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238aa43c836b59fdaefdc759e10e595aa562b4aed347224117387b7e8c4e4cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b92be065d943e991341bdf58d435913fc5695655686a16101abb5c41f8d3ecc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.995849 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.995981 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.996002 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.996044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.996070 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:24Z","lastTransitionTime":"2026-01-22T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:24 crc kubenswrapper[4800]: I0122 00:08:24.999331 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbed54756026c0ad5c9d83988118ce6ca9c1d6403cdd9a8acfcacb8dbe43bfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:24Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.023563 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bbdee4d3753786cb923a82dd53217b526d55c8e16fe5dd9a929483366c209a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.047207 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.066595 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fp7s5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b37dab4b-08e1-4daa-86d9-561432a82428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f712889a0d454c604047757381e1977c7fe40bcab5ddf2af512d47503156817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45ldt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fp7s5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.092935 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vfgd4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aff2b22c-b87d-47c9-b5a0-6e5772022488\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:09Z\\\",\\\"message\\\":\\\"2026-01-22T00:07:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb\\\\n2026-01-22T00:07:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c0d8b6c3-328e-45df-bd32-8f7c47cb32cb to /host/opt/cni/bin/\\\\n2026-01-22T00:07:24Z [verbose] multus-daemon started\\\\n2026-01-22T00:07:24Z [verbose] Readiness Indicator file check\\\\n2026-01-22T00:08:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sjcf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vfgd4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.099578 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.099644 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.099663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.099694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.099716 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.120983 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46f4c09a-7f56-40ab-b92f-6e01c949a38e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T00:08:17Z\\\",\\\"message\\\":\\\"ocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 00:08:16.875061 6809 lb_config.go:1031] Cluster endpoints for openshift-console-operator/metrics for network=default are: map[]\\\\nI0122 00:08:16.874822 6809 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI0122 00:08:16.874838 6809 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fp7s5\\\\nI0122 00:08:16.875062 6809 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0122 00:08:16.874968 6809 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0122 00:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:08:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gx9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w4dww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.143552 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89879ce3-b327-4736-9547-c2ff3c97a5d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d09764a1c212569f84dc46c9eeda1d26917ff0cc7ab2d61ceb3e4f088d015eda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03686dd9c9682d8502c4d699e03678b55b57290abb1206d3fb73a61b13539fd0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://851b6246162caceae3d3016a4352e4887549e7ce3c394d6326e28c9d0115431f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.173497 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\":120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 00:07:22.618373 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 00:07:22.618398 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618403 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 00:07:22.618408 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 00:07:22.618411 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 00:07:22.618414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 00:07:22.618417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 00:07:22.618612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0122 00:07:22.622017 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769040426\\\\\\\\\\\\\\\" (2026-01-22 00:07:06 +0000 UTC to 2026-02-21 00:07:07 +0000 UTC (now=2026-01-22 00:07:22.621976238 +0000 UTC))\\\\\\\"\\\\nI0122 00:07:22.622063 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622138 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0122 00:07:22.622158 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2043988252/tls.crt::/tmp/serving-cert-2043988252/tls.key\\\\\\\"\\\\nI0122 00:07:22.622199 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nF0122 00:07:22.622196 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.191202 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxlxs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.202991 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.203054 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.203077 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.203101 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.203123 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.209757 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.225810 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9abc3993-c3c9-4b93-a963-13dee17b8cff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T00:07:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a11221698e6fc7c55859b863a17d52dc0f5f46d8c1417a59995cb81caf7a024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2688b9212bec561edd8ba5d87ceaff43df5824a1cfd27cc41c5919807adb15ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b2bm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T00:07:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7zgrm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T00:08:25Z is after 2025-08-24T17:21:41Z" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.306529 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.306591 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.306602 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.306621 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.306632 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.410072 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.410130 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.410144 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.410166 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.410184 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.513565 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.514037 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.514184 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.514354 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.514512 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.619715 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.619792 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.619817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.619855 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.619881 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.723172 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.723242 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.723262 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.723291 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.723310 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.825534 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 05:49:10.004015712 +0000 UTC Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.827207 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.827286 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.827302 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.827325 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.827379 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.931686 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.931740 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.931755 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.931774 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:25 crc kubenswrapper[4800]: I0122 00:08:25.931786 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:25Z","lastTransitionTime":"2026-01-22T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.035430 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.035488 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.035504 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.035526 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.035541 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.139770 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.139853 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.139932 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.140003 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.140029 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.244044 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.244161 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.244180 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.244211 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.244238 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.348071 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.348154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.348185 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.348224 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.348252 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.461140 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.461250 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.461268 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.461305 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.461323 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.565228 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.565319 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.565347 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.565389 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.565415 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.668679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.669031 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.669056 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.669090 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.669116 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.772742 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.773220 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.773246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.773278 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.773298 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.817355 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.817355 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.817448 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.817593 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.817725 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.818069 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.818116 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.818386 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.826743 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 07:02:02.421087905 +0000 UTC Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.829354 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.829585 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.829545664 +0000 UTC m=+146.309848832 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.829655 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.829742 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.829946 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.829984 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830008 4800 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830071 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830104 4800 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830126 4800 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830085 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.830069836 +0000 UTC m=+146.310372914 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830220 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.830192739 +0000 UTC m=+146.310495817 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.830294 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.830392 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830499 4800 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830600 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.830573548 +0000 UTC m=+146.310876596 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830652 4800 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: E0122 00:08:26.830720 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.830701741 +0000 UTC m=+146.311004819 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.876801 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.876880 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.876935 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.876962 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.876985 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.979938 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.979992 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.980022 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.980040 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:26 crc kubenswrapper[4800]: I0122 00:08:26.980050 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:26Z","lastTransitionTime":"2026-01-22T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.084159 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.084238 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.084250 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.084271 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.084284 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.187523 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.187617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.187641 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.187677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.187705 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.290789 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.290847 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.290863 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.292109 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.292152 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.395634 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.395694 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.395705 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.395726 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.395742 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.498766 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.498821 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.498837 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.498862 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.498880 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.601629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.601690 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.601712 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.601736 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.601751 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.704935 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.705005 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.705028 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.705065 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.705089 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.808351 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.808424 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.808442 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.808476 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.808497 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.827231 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 05:53:22.313354877 +0000 UTC Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.912264 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.912337 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.912360 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.912390 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:27 crc kubenswrapper[4800]: I0122 00:08:27.912411 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:27Z","lastTransitionTime":"2026-01-22T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.016460 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.016556 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.016575 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.016608 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.016636 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.121060 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.121147 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.121172 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.121208 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.121230 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.225600 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.225679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.225699 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.225729 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.225750 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.328763 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.328849 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.328869 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.328936 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.328958 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.432123 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.432225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.432254 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.432293 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.432317 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.535818 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.535920 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.535967 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.536006 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.536049 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.640024 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.640109 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.640127 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.640156 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.640177 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.743162 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.743247 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.743267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.743298 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.743318 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.817562 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.817714 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:28 crc kubenswrapper[4800]: E0122 00:08:28.817811 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.817839 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:28 crc kubenswrapper[4800]: E0122 00:08:28.818088 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.818155 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:28 crc kubenswrapper[4800]: E0122 00:08:28.818285 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:28 crc kubenswrapper[4800]: E0122 00:08:28.818466 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.828036 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 04:22:10.035143752 +0000 UTC Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.846868 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.846962 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.846983 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.847010 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.847030 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.950867 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.950978 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.951003 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.951037 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:28 crc kubenswrapper[4800]: I0122 00:08:28.951058 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:28Z","lastTransitionTime":"2026-01-22T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.054817 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.054881 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.054929 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.054958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.054978 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.159459 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.159539 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.159559 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.159589 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.159610 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.263421 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.263494 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.263511 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.263542 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.263561 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.366775 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.366849 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.366870 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.366939 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.366960 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.470731 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.470797 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.470807 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.470842 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.470856 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.574304 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.574392 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.574417 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.574452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.574476 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.678184 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.678326 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.678345 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.678375 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.678394 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.781629 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.781686 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.781697 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.781724 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.781738 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.819548 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:08:29 crc kubenswrapper[4800]: E0122 00:08:29.821354 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.828598 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 22:42:20.110595164 +0000 UTC Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.885769 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.885833 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.885854 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.885921 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.885947 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.991715 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.991756 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.991766 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.991784 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:29 crc kubenswrapper[4800]: I0122 00:08:29.991795 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:29Z","lastTransitionTime":"2026-01-22T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.095679 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.096085 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.096157 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.096225 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.096286 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.199617 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.199672 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.199688 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.199714 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.199733 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.303093 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.303181 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.303200 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.303236 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.303258 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.406537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.406985 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.407233 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.407477 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.407721 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.511095 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.511184 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.511209 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.511247 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.511274 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.614755 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.614865 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.614919 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.614950 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.614970 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.718075 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.718128 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.718138 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.718156 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.718175 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.817786 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.817981 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:30 crc kubenswrapper[4800]: E0122 00:08:30.818068 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.818127 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.818245 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:30 crc kubenswrapper[4800]: E0122 00:08:30.818423 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:30 crc kubenswrapper[4800]: E0122 00:08:30.818506 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:30 crc kubenswrapper[4800]: E0122 00:08:30.818617 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.820505 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.820557 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.820577 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.820604 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.820628 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.829313 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 20:25:54.114085008 +0000 UTC Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.923708 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.924228 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.924369 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.924516 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:30 crc kubenswrapper[4800]: I0122 00:08:30.924670 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:30Z","lastTransitionTime":"2026-01-22T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.028403 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.028930 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.029139 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.029360 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.029543 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.133386 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.133447 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.133460 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.133486 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.133503 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.237717 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.237805 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.237827 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.237860 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.237880 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.340721 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.340791 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.340830 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.340866 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.340928 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.444814 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.444934 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.444958 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.444990 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.445014 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.547800 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.547942 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.547972 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.548017 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.548041 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.652267 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.652339 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.652368 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.652400 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.652484 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.756029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.756107 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.756135 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.756167 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.756188 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.829853 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 11:02:16.687007347 +0000 UTC Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.859873 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.859966 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.859983 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.860029 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.860052 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.965404 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.965491 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.965517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.965549 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:31 crc kubenswrapper[4800]: I0122 00:08:31.965574 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:31Z","lastTransitionTime":"2026-01-22T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.069551 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.069630 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.069649 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.069677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.069698 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.173586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.173665 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.173683 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.174255 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.174313 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.277762 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.277835 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.277852 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.277879 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.277927 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.381684 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.381773 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.381794 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.381823 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.381846 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.484627 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.484668 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.484677 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.484692 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.484702 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.587384 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.587432 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.587443 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.587459 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.587470 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.690231 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.690271 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.690281 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.690298 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.690309 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.793239 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.793299 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.793316 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.793343 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.793385 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.818086 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.818157 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:32 crc kubenswrapper[4800]: E0122 00:08:32.818267 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.818101 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:32 crc kubenswrapper[4800]: E0122 00:08:32.818411 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:32 crc kubenswrapper[4800]: E0122 00:08:32.818496 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.818566 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:32 crc kubenswrapper[4800]: E0122 00:08:32.818639 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.830781 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 21:46:58.546267527 +0000 UTC Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.897349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.897412 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.897432 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.897461 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:32 crc kubenswrapper[4800]: I0122 00:08:32.897480 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:32Z","lastTransitionTime":"2026-01-22T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.000318 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.000363 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.000373 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.000394 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.000409 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.104276 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.104331 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.104349 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.104376 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.104396 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.207572 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.207663 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.207680 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.207708 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.207728 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.311281 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.311335 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.311348 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.311369 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.311383 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.415041 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.415127 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.415146 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.415176 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.415197 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.518114 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.518186 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.518212 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.518246 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.518273 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.621592 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.621674 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.621700 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.621736 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.621763 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.725224 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.725306 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.725332 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.725365 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.725391 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.829437 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.829507 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.829531 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.829563 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.829586 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.831737 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 20:21:51.693585024 +0000 UTC Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.932326 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.932401 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.932428 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.932456 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:33 crc kubenswrapper[4800]: I0122 00:08:33.932479 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:33Z","lastTransitionTime":"2026-01-22T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.035154 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.035216 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.035234 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.035258 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.035274 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.138452 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.138527 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.138553 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.138586 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.138608 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.241716 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.241789 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.241809 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.241844 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.241863 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.345369 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.345444 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.345481 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.345515 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.345536 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.449425 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.449517 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.449537 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.449569 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.449588 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.555610 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.555728 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.555748 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.555780 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.555804 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.660775 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.660858 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.660882 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.660955 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.660992 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.765023 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.765094 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.765118 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.765151 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.765172 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.817477 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:34 crc kubenswrapper[4800]: E0122 00:08:34.817711 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.817843 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.817978 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:34 crc kubenswrapper[4800]: E0122 00:08:34.818120 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:34 crc kubenswrapper[4800]: E0122 00:08:34.818315 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.818478 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:34 crc kubenswrapper[4800]: E0122 00:08:34.818625 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.833630 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 06:10:19.803168017 +0000 UTC Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.860799 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7tj9m" podStartSLOduration=72.860767741 podStartE2EDuration="1m12.860767741s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:34.860404233 +0000 UTC m=+90.340707311" watchObservedRunningTime="2026-01-22 00:08:34.860767741 +0000 UTC m=+90.341070809" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.870070 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.870153 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.870179 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.870214 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.870238 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.955515 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podStartSLOduration=72.955484723 podStartE2EDuration="1m12.955484723s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:34.955380849 +0000 UTC m=+90.435683897" watchObservedRunningTime="2026-01-22 00:08:34.955484723 +0000 UTC m=+90.435787801" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.955817 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-bv9wt" podStartSLOduration=72.95580113 podStartE2EDuration="1m12.95580113s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:34.935376122 +0000 UTC m=+90.415679180" watchObservedRunningTime="2026-01-22 00:08:34.95580113 +0000 UTC m=+90.436104208" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.973950 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.974033 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.974054 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.974081 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:34 crc kubenswrapper[4800]: I0122 00:08:34.974108 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:34Z","lastTransitionTime":"2026-01-22T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.002381 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-fp7s5" podStartSLOduration=73.002336571 podStartE2EDuration="1m13.002336571s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:34.977083698 +0000 UTC m=+90.457386746" watchObservedRunningTime="2026-01-22 00:08:35.002336571 +0000 UTC m=+90.482639649" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.047119 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vfgd4" podStartSLOduration=73.047079589 podStartE2EDuration="1m13.047079589s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.003848907 +0000 UTC m=+90.484151955" watchObservedRunningTime="2026-01-22 00:08:35.047079589 +0000 UTC m=+90.527382667" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.077348 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.077399 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.077415 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.077451 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.077468 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:35Z","lastTransitionTime":"2026-01-22T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.096601 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.09656955 podStartE2EDuration="1m9.09656955s" podCreationTimestamp="2026-01-22 00:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.07098899 +0000 UTC m=+90.551292058" watchObservedRunningTime="2026-01-22 00:08:35.09656955 +0000 UTC m=+90.576872598" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.097089 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.097084223 podStartE2EDuration="1m13.097084223s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.09695813 +0000 UTC m=+90.577261178" watchObservedRunningTime="2026-01-22 00:08:35.097084223 +0000 UTC m=+90.577387271" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.116233 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.116318 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.116341 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.116414 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.116443 4800 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T00:08:35Z","lastTransitionTime":"2026-01-22T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.184784 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7zgrm" podStartSLOduration=72.184757285 podStartE2EDuration="1m12.184757285s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.183427263 +0000 UTC m=+90.663730301" watchObservedRunningTime="2026-01-22 00:08:35.184757285 +0000 UTC m=+90.665060343" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.187608 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9"] Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.188190 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.189831 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.190317 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.191376 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.202157 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.239414 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.239468 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9018277-4c85-4233-944b-cc40d91c7b85-service-ca\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.239493 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9018277-4c85-4233-944b-cc40d91c7b85-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.239509 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.239543 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9018277-4c85-4233-944b-cc40d91c7b85-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.251257 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.251237913 podStartE2EDuration="1m8.251237913s" podCreationTimestamp="2026-01-22 00:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.25073683 +0000 UTC m=+90.731039868" watchObservedRunningTime="2026-01-22 00:08:35.251237913 +0000 UTC m=+90.731540951" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.251755 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.251750105 podStartE2EDuration="17.251750105s" podCreationTimestamp="2026-01-22 00:08:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.214513536 +0000 UTC m=+90.694816574" watchObservedRunningTime="2026-01-22 00:08:35.251750105 +0000 UTC m=+90.732053143" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.264625 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=39.264600161 podStartE2EDuration="39.264600161s" podCreationTimestamp="2026-01-22 00:07:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:35.264370226 +0000 UTC m=+90.744673264" watchObservedRunningTime="2026-01-22 00:08:35.264600161 +0000 UTC m=+90.744903199" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341054 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9018277-4c85-4233-944b-cc40d91c7b85-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341116 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341144 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9018277-4c85-4233-944b-cc40d91c7b85-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341228 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341253 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9018277-4c85-4233-944b-cc40d91c7b85-service-ca\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341271 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.341408 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9018277-4c85-4233-944b-cc40d91c7b85-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.342448 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9018277-4c85-4233-944b-cc40d91c7b85-service-ca\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.348274 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9018277-4c85-4233-944b-cc40d91c7b85-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.357512 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9018277-4c85-4233-944b-cc40d91c7b85-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-snlp9\" (UID: \"c9018277-4c85-4233-944b-cc40d91c7b85\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.513851 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.834575 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 00:47:29.497512814 +0000 UTC Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.835077 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 22 00:08:35 crc kubenswrapper[4800]: I0122 00:08:35.844944 4800 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.506809 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" event={"ID":"c9018277-4c85-4233-944b-cc40d91c7b85","Type":"ContainerStarted","Data":"2d74d0f9459f4a9a4f58afabb284e9a911b16dfb7f37108f668ce43e8300fb4c"} Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.506879 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" event={"ID":"c9018277-4c85-4233-944b-cc40d91c7b85","Type":"ContainerStarted","Data":"73f8d95e0d4098ad7d39c17932fc74b04e66b254412cb89799a5b87a842f955a"} Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.818233 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.818349 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.818418 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:36 crc kubenswrapper[4800]: I0122 00:08:36.818470 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:36 crc kubenswrapper[4800]: E0122 00:08:36.818614 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:36 crc kubenswrapper[4800]: E0122 00:08:36.818794 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:36 crc kubenswrapper[4800]: E0122 00:08:36.819280 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:36 crc kubenswrapper[4800]: E0122 00:08:36.819505 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:38 crc kubenswrapper[4800]: I0122 00:08:38.817971 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:38 crc kubenswrapper[4800]: E0122 00:08:38.818149 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:38 crc kubenswrapper[4800]: I0122 00:08:38.818278 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:38 crc kubenswrapper[4800]: I0122 00:08:38.818341 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:38 crc kubenswrapper[4800]: I0122 00:08:38.818255 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:38 crc kubenswrapper[4800]: E0122 00:08:38.818488 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:38 crc kubenswrapper[4800]: E0122 00:08:38.818728 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:38 crc kubenswrapper[4800]: E0122 00:08:38.818781 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:40 crc kubenswrapper[4800]: I0122 00:08:40.818171 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:40 crc kubenswrapper[4800]: I0122 00:08:40.818214 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:40 crc kubenswrapper[4800]: I0122 00:08:40.818302 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:40 crc kubenswrapper[4800]: I0122 00:08:40.818331 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:40 crc kubenswrapper[4800]: E0122 00:08:40.818434 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:40 crc kubenswrapper[4800]: E0122 00:08:40.818846 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:40 crc kubenswrapper[4800]: E0122 00:08:40.819200 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:40 crc kubenswrapper[4800]: E0122 00:08:40.819490 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:41 crc kubenswrapper[4800]: I0122 00:08:41.637766 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:41 crc kubenswrapper[4800]: E0122 00:08:41.638021 4800 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:08:41 crc kubenswrapper[4800]: E0122 00:08:41.638101 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs podName:ca19c294-fff4-4fe6-a1a4-25e04d0624b3 nodeName:}" failed. No retries permitted until 2026-01-22 00:09:45.638078262 +0000 UTC m=+161.118381320 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs") pod "network-metrics-daemon-svkb9" (UID: "ca19c294-fff4-4fe6-a1a4-25e04d0624b3") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 00:08:42 crc kubenswrapper[4800]: I0122 00:08:42.818216 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:42 crc kubenswrapper[4800]: I0122 00:08:42.818357 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:42 crc kubenswrapper[4800]: I0122 00:08:42.818366 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:42 crc kubenswrapper[4800]: E0122 00:08:42.818595 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:42 crc kubenswrapper[4800]: I0122 00:08:42.818633 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:42 crc kubenswrapper[4800]: E0122 00:08:42.818723 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:42 crc kubenswrapper[4800]: E0122 00:08:42.818827 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:42 crc kubenswrapper[4800]: E0122 00:08:42.819031 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:44 crc kubenswrapper[4800]: I0122 00:08:44.817142 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:44 crc kubenswrapper[4800]: I0122 00:08:44.817258 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:44 crc kubenswrapper[4800]: I0122 00:08:44.817157 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:44 crc kubenswrapper[4800]: E0122 00:08:44.817345 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:44 crc kubenswrapper[4800]: E0122 00:08:44.819519 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:44 crc kubenswrapper[4800]: I0122 00:08:44.819574 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:44 crc kubenswrapper[4800]: E0122 00:08:44.820244 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:44 crc kubenswrapper[4800]: E0122 00:08:44.820347 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:44 crc kubenswrapper[4800]: I0122 00:08:44.820547 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:08:44 crc kubenswrapper[4800]: E0122 00:08:44.820738 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:46 crc kubenswrapper[4800]: I0122 00:08:46.817989 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:46 crc kubenswrapper[4800]: I0122 00:08:46.818050 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:46 crc kubenswrapper[4800]: I0122 00:08:46.818085 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:46 crc kubenswrapper[4800]: I0122 00:08:46.818011 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:46 crc kubenswrapper[4800]: E0122 00:08:46.818141 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:46 crc kubenswrapper[4800]: E0122 00:08:46.818241 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:46 crc kubenswrapper[4800]: E0122 00:08:46.818314 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:46 crc kubenswrapper[4800]: E0122 00:08:46.818364 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:48 crc kubenswrapper[4800]: I0122 00:08:48.817933 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:48 crc kubenswrapper[4800]: I0122 00:08:48.818032 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:48 crc kubenswrapper[4800]: E0122 00:08:48.818078 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:48 crc kubenswrapper[4800]: E0122 00:08:48.818199 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:48 crc kubenswrapper[4800]: I0122 00:08:48.818347 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:48 crc kubenswrapper[4800]: I0122 00:08:48.818382 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:48 crc kubenswrapper[4800]: E0122 00:08:48.818546 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:48 crc kubenswrapper[4800]: E0122 00:08:48.818676 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:50 crc kubenswrapper[4800]: I0122 00:08:50.818113 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:50 crc kubenswrapper[4800]: I0122 00:08:50.818189 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:50 crc kubenswrapper[4800]: I0122 00:08:50.818238 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:50 crc kubenswrapper[4800]: E0122 00:08:50.818352 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:50 crc kubenswrapper[4800]: I0122 00:08:50.818113 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:50 crc kubenswrapper[4800]: E0122 00:08:50.818546 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:50 crc kubenswrapper[4800]: E0122 00:08:50.818610 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:50 crc kubenswrapper[4800]: E0122 00:08:50.818651 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:52 crc kubenswrapper[4800]: I0122 00:08:52.817852 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:52 crc kubenswrapper[4800]: I0122 00:08:52.817992 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:52 crc kubenswrapper[4800]: I0122 00:08:52.817870 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:52 crc kubenswrapper[4800]: E0122 00:08:52.818164 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:52 crc kubenswrapper[4800]: I0122 00:08:52.818340 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:52 crc kubenswrapper[4800]: E0122 00:08:52.818336 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:52 crc kubenswrapper[4800]: E0122 00:08:52.818503 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:52 crc kubenswrapper[4800]: E0122 00:08:52.818543 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:54 crc kubenswrapper[4800]: I0122 00:08:54.817699 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:54 crc kubenswrapper[4800]: I0122 00:08:54.817722 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:54 crc kubenswrapper[4800]: I0122 00:08:54.817839 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:54 crc kubenswrapper[4800]: I0122 00:08:54.817957 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:54 crc kubenswrapper[4800]: E0122 00:08:54.819404 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:54 crc kubenswrapper[4800]: E0122 00:08:54.819640 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:54 crc kubenswrapper[4800]: E0122 00:08:54.819777 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:54 crc kubenswrapper[4800]: E0122 00:08:54.819741 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.577854 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/1.log" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.578572 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/0.log" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.578618 4800 generic.go:334] "Generic (PLEG): container finished" podID="aff2b22c-b87d-47c9-b5a0-6e5772022488" containerID="7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30" exitCode=1 Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.578656 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerDied","Data":"7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30"} Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.578702 4800 scope.go:117] "RemoveContainer" containerID="7d1c910a2b5be0b6272c2a8a4c6073c52778edd5a8ce18a8ce84862a9fecb35a" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.580195 4800 scope.go:117] "RemoveContainer" containerID="7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.580512 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vfgd4_openshift-multus(aff2b22c-b87d-47c9-b5a0-6e5772022488)\"" pod="openshift-multus/multus-vfgd4" podUID="aff2b22c-b87d-47c9-b5a0-6e5772022488" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.603056 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-snlp9" podStartSLOduration=94.603027365 podStartE2EDuration="1m34.603027365s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:08:36.538602405 +0000 UTC m=+92.018905483" watchObservedRunningTime="2026-01-22 00:08:56.603027365 +0000 UTC m=+112.083330433" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.817413 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.817473 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.817500 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.817413 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.817655 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.817836 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.818280 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.818402 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:56 crc kubenswrapper[4800]: I0122 00:08:56.818760 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:08:56 crc kubenswrapper[4800]: E0122 00:08:56.818998 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w4dww_openshift-ovn-kubernetes(46f4c09a-7f56-40ab-b92f-6e01c949a38e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" Jan 22 00:08:57 crc kubenswrapper[4800]: I0122 00:08:57.583143 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/1.log" Jan 22 00:08:58 crc kubenswrapper[4800]: I0122 00:08:58.817513 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:08:58 crc kubenswrapper[4800]: I0122 00:08:58.817652 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:08:58 crc kubenswrapper[4800]: I0122 00:08:58.817565 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:08:58 crc kubenswrapper[4800]: I0122 00:08:58.817572 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:08:58 crc kubenswrapper[4800]: E0122 00:08:58.817812 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:08:58 crc kubenswrapper[4800]: E0122 00:08:58.818010 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:08:58 crc kubenswrapper[4800]: E0122 00:08:58.818159 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:08:58 crc kubenswrapper[4800]: E0122 00:08:58.818304 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:00 crc kubenswrapper[4800]: I0122 00:09:00.817715 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:00 crc kubenswrapper[4800]: I0122 00:09:00.817856 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:00 crc kubenswrapper[4800]: E0122 00:09:00.818411 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:00 crc kubenswrapper[4800]: I0122 00:09:00.818061 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:00 crc kubenswrapper[4800]: E0122 00:09:00.818585 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:00 crc kubenswrapper[4800]: I0122 00:09:00.817867 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:00 crc kubenswrapper[4800]: E0122 00:09:00.818804 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:00 crc kubenswrapper[4800]: E0122 00:09:00.818979 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:02 crc kubenswrapper[4800]: I0122 00:09:02.817782 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:02 crc kubenswrapper[4800]: I0122 00:09:02.818236 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:02 crc kubenswrapper[4800]: I0122 00:09:02.818414 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:02 crc kubenswrapper[4800]: I0122 00:09:02.818541 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:02 crc kubenswrapper[4800]: E0122 00:09:02.818502 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:02 crc kubenswrapper[4800]: E0122 00:09:02.818645 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:02 crc kubenswrapper[4800]: E0122 00:09:02.818950 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:02 crc kubenswrapper[4800]: E0122 00:09:02.819685 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.764305 4800 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 22 00:09:04 crc kubenswrapper[4800]: I0122 00:09:04.817629 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:04 crc kubenswrapper[4800]: I0122 00:09:04.817677 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.818504 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:04 crc kubenswrapper[4800]: I0122 00:09:04.818582 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:04 crc kubenswrapper[4800]: I0122 00:09:04.818685 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.818728 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.818881 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.819001 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:04 crc kubenswrapper[4800]: E0122 00:09:04.915421 4800 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 00:09:06 crc kubenswrapper[4800]: I0122 00:09:06.817269 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:06 crc kubenswrapper[4800]: I0122 00:09:06.817419 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:06 crc kubenswrapper[4800]: I0122 00:09:06.817511 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:06 crc kubenswrapper[4800]: E0122 00:09:06.817499 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:06 crc kubenswrapper[4800]: I0122 00:09:06.817596 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:06 crc kubenswrapper[4800]: E0122 00:09:06.817742 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:06 crc kubenswrapper[4800]: E0122 00:09:06.817877 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:06 crc kubenswrapper[4800]: E0122 00:09:06.818095 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:08 crc kubenswrapper[4800]: I0122 00:09:08.818040 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:08 crc kubenswrapper[4800]: I0122 00:09:08.818085 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:08 crc kubenswrapper[4800]: I0122 00:09:08.818098 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:08 crc kubenswrapper[4800]: I0122 00:09:08.818422 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:08 crc kubenswrapper[4800]: E0122 00:09:08.818586 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:08 crc kubenswrapper[4800]: E0122 00:09:08.818624 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:08 crc kubenswrapper[4800]: E0122 00:09:08.818771 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:08 crc kubenswrapper[4800]: E0122 00:09:08.818378 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:08 crc kubenswrapper[4800]: I0122 00:09:08.818989 4800 scope.go:117] "RemoveContainer" containerID="7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30" Jan 22 00:09:09 crc kubenswrapper[4800]: I0122 00:09:09.626969 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/1.log" Jan 22 00:09:09 crc kubenswrapper[4800]: I0122 00:09:09.627039 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerStarted","Data":"d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf"} Jan 22 00:09:09 crc kubenswrapper[4800]: E0122 00:09:09.917280 4800 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 00:09:10 crc kubenswrapper[4800]: I0122 00:09:10.817720 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:10 crc kubenswrapper[4800]: I0122 00:09:10.817838 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:10 crc kubenswrapper[4800]: I0122 00:09:10.817940 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:10 crc kubenswrapper[4800]: E0122 00:09:10.817928 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:10 crc kubenswrapper[4800]: I0122 00:09:10.817953 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:10 crc kubenswrapper[4800]: E0122 00:09:10.818102 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:10 crc kubenswrapper[4800]: E0122 00:09:10.818310 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:10 crc kubenswrapper[4800]: E0122 00:09:10.818439 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:11 crc kubenswrapper[4800]: I0122 00:09:11.818664 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.641200 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/3.log" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.644367 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerStarted","Data":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.644920 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.676532 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podStartSLOduration=109.676501026 podStartE2EDuration="1m49.676501026s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:12.673931718 +0000 UTC m=+128.154234756" watchObservedRunningTime="2026-01-22 00:09:12.676501026 +0000 UTC m=+128.156804104" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.753148 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-svkb9"] Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.753359 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:12 crc kubenswrapper[4800]: E0122 00:09:12.753531 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.817652 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.817684 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:12 crc kubenswrapper[4800]: E0122 00:09:12.817777 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:12 crc kubenswrapper[4800]: I0122 00:09:12.817955 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:12 crc kubenswrapper[4800]: E0122 00:09:12.818075 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:12 crc kubenswrapper[4800]: E0122 00:09:12.818294 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:14 crc kubenswrapper[4800]: I0122 00:09:14.818225 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:14 crc kubenswrapper[4800]: I0122 00:09:14.818252 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:14 crc kubenswrapper[4800]: E0122 00:09:14.820985 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:14 crc kubenswrapper[4800]: I0122 00:09:14.821052 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:14 crc kubenswrapper[4800]: E0122 00:09:14.821239 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:14 crc kubenswrapper[4800]: I0122 00:09:14.821326 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:14 crc kubenswrapper[4800]: E0122 00:09:14.821386 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:14 crc kubenswrapper[4800]: E0122 00:09:14.821552 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:14 crc kubenswrapper[4800]: E0122 00:09:14.918280 4800 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 00:09:16 crc kubenswrapper[4800]: I0122 00:09:16.817864 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:16 crc kubenswrapper[4800]: I0122 00:09:16.818011 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:16 crc kubenswrapper[4800]: I0122 00:09:16.818118 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:16 crc kubenswrapper[4800]: E0122 00:09:16.818022 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:16 crc kubenswrapper[4800]: E0122 00:09:16.818214 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:16 crc kubenswrapper[4800]: I0122 00:09:16.818315 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:16 crc kubenswrapper[4800]: E0122 00:09:16.818491 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:16 crc kubenswrapper[4800]: E0122 00:09:16.818817 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:18 crc kubenswrapper[4800]: I0122 00:09:18.818093 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:18 crc kubenswrapper[4800]: I0122 00:09:18.818161 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:18 crc kubenswrapper[4800]: I0122 00:09:18.818112 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:18 crc kubenswrapper[4800]: E0122 00:09:18.818346 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 00:09:18 crc kubenswrapper[4800]: I0122 00:09:18.818378 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:18 crc kubenswrapper[4800]: E0122 00:09:18.818501 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 00:09:18 crc kubenswrapper[4800]: E0122 00:09:18.818610 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 00:09:18 crc kubenswrapper[4800]: E0122 00:09:18.818761 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svkb9" podUID="ca19c294-fff4-4fe6-a1a4-25e04d0624b3" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.817568 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.817974 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.818206 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.818295 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.825139 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.825385 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.825422 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.825485 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.825733 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 00:09:20 crc kubenswrapper[4800]: I0122 00:09:20.826722 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.291186 4800 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.347998 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.348468 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.357006 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.357417 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.357435 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.357513 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.357868 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.359048 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x99cx"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.360282 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.360475 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hbblt"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.361287 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.375113 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.375134 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29484000-kr6nl"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.375973 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.376331 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.376488 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.376557 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6t2gz"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.376620 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.377058 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.377250 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.378368 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.379870 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.380148 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.380438 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.380609 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.381366 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.381492 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-shzt2"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.387085 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.388047 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.388245 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.388558 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.388658 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.391065 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.391208 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.391630 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.391914 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.392920 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ggnbw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.393285 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.393638 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tpn7g"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.394282 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.395591 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.396137 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.396150 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-q6r6s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.397494 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.399146 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.399333 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.401002 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.428315 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq4b8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.446785 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.447596 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.451608 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.451927 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.452173 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zn8zh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.452745 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.453053 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.453223 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.453697 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.453809 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.453238 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454014 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454160 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454278 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454362 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454445 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454530 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454617 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454639 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454694 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454783 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454870 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454945 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.454978 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455241 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455387 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455465 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455404 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455582 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.455671 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.459060 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.459806 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cv9lb"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460307 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460345 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460364 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-client\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460375 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460510 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460382 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-config\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460637 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460652 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460682 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-service-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460727 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcxq5\" (UniqueName: \"kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460793 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.460897 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461057 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhj5\" (UniqueName: \"kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461161 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461253 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-images\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461360 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461405 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461469 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461504 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-serving-cert\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461538 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbzg4\" (UniqueName: \"kubernetes.io/projected/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-kube-api-access-nbzg4\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461620 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-config\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461658 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461694 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wd2f\" (UniqueName: \"kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461753 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx9fq\" (UniqueName: \"kubernetes.io/projected/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-kube-api-access-wx9fq\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.461763 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462082 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462124 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462235 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462297 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462402 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462464 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462493 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462586 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462636 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462706 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462724 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462809 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462845 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462950 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463026 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463044 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463137 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463263 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.462811 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463371 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463417 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463482 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.463742 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.464156 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.475171 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.476627 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.477091 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.478039 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.478375 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.478560 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.479302 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.479428 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.479502 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480131 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480281 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480404 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480564 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480659 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.480733 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.481873 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.482354 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.482438 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.486036 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.489767 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.492119 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.492549 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.492626 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.492936 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.504585 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.504829 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.506244 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.508404 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.509929 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.516980 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.517520 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.527019 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.531351 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.534369 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.534691 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.535186 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ffvfb"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.535595 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.535825 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.538512 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.538737 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.539016 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.539013 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.540019 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.541071 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.541428 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.544287 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.545370 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.545500 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.546419 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.547735 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.549053 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.549444 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.552563 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wv42v"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.552664 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.553502 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.553536 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.554723 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.555129 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.556259 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.557358 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-42dwh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.557863 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.558018 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.558191 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.558390 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.561725 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562525 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcxq5\" (UniqueName: \"kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562567 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562597 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhj5\" (UniqueName: \"kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562626 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562646 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562680 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-images\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562710 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562731 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562753 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562776 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-serving-cert\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562797 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbzg4\" (UniqueName: \"kubernetes.io/projected/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-kube-api-access-nbzg4\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562830 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-config\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562855 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562879 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wd2f\" (UniqueName: \"kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562930 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx9fq\" (UniqueName: \"kubernetes.io/projected/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-kube-api-access-wx9fq\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562968 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.562994 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.563021 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-client\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.563044 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-config\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.563081 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.563105 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-service-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.564281 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.564751 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.565036 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.565091 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.565621 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.565848 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566097 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-config\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566282 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-images\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566473 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566537 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566713 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566801 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.566807 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.567052 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-config\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.567078 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-service-ca\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.567163 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.567721 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.568240 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.569469 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.570431 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.570492 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.570971 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.573158 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.579624 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29484000-kr6nl"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.581805 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.594008 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ggnbw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.594094 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x99cx"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.594106 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6t2gz"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.599154 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.599344 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.599674 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.599769 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-etcd-client\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.599410 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.604498 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.604681 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.604772 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-serving-cert\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.605921 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.608534 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q6r6s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.609901 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.610084 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.611271 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-shzt2"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.612527 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.613721 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6l6mb"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.615287 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.615417 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.616149 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tpn7g"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.617394 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.618757 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.619767 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.620778 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.621779 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.623876 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.625282 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hbblt"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.626431 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wv42v"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.627240 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.629785 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.630018 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.630416 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.631682 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.633457 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.634261 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4hl4v"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.635684 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmzt"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.636713 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.637044 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq4b8"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.637134 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.638661 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.638766 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.640861 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.642591 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6l6mb"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.644237 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.645277 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cv9lb"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.646304 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmzt"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.648009 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.648996 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zn8zh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.650438 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-42dwh"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.650570 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.651496 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.652592 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.653616 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-g5tbw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.654571 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.655005 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g5tbw"] Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.669942 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.691151 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.711291 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.730393 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.751749 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.771424 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.791723 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.811694 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.830719 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.865809 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.874491 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.890748 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.909990 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.930211 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.951711 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.970843 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 00:09:26 crc kubenswrapper[4800]: I0122 00:09:26.991081 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.011754 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.050760 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.071345 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.092265 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.111862 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.130867 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.151011 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.171816 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.191769 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.213860 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.231068 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.252417 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.272675 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.291028 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.312723 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.332229 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.351103 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.373572 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.412529 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.430827 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439015 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439482 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-trusted-ca-bundle\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439680 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439746 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439778 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439801 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f89rt\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439823 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439869 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.439914 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.440096 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.440198 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.440235 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-service-ca\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.440287 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-oauth-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.440832 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:27.940809266 +0000 UTC m=+143.421112314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.450872 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.471610 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.493192 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.519178 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.532165 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.541530 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.541791 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.041751675 +0000 UTC m=+143.522054703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542302 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce65736c-5fd3-4d03-9cf2-e8759f560551-trusted-ca\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542446 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542556 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542652 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czrzm\" (UniqueName: \"kubernetes.io/projected/b081f2a5-0dee-4ee9-b5af-564af14f3e69-kube-api-access-czrzm\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542764 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znntx\" (UniqueName: \"kubernetes.io/projected/d516372c-17a3-438c-87f8-d2e3edfb5670-kube-api-access-znntx\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.542950 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hkrs\" (UniqueName: \"kubernetes.io/projected/c86f998b-7292-4a18-a830-7d2c46dcdf2c-kube-api-access-5hkrs\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543059 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-client\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543189 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-serving-cert\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543311 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/029287ae-ef37-46c9-b023-1d7074803fd6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543431 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6de60db-1a8c-4be4-9943-c5afcd8f4566-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543620 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v65qb\" (UniqueName: \"kubernetes.io/projected/f017341d-0ebc-442d-8980-5e2839c89c3a-kube-api-access-v65qb\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543751 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.543957 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-node-pullsecrets\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544112 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f017341d-0ebc-442d-8980-5e2839c89c3a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544209 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e26d7b-9ede-44a2-97d6-f60b80b38839-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544323 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gkd\" (UniqueName: \"kubernetes.io/projected/a5e37c96-49b2-4450-898d-d01847863fff-kube-api-access-57gkd\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544373 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88jt8\" (UniqueName: \"kubernetes.io/projected/092d47e3-8067-4702-bbd2-1a1bdb429b67-kube-api-access-88jt8\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544434 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-plugins-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544535 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544666 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544783 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-dir\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.544934 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzwkz\" (UniqueName: \"kubernetes.io/projected/dbeac130-c061-4ae1-a609-8808eb849e38-kube-api-access-gzwkz\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545076 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e26d7b-9ede-44a2-97d6-f60b80b38839-config\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545189 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-trusted-ca-bundle\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545301 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2zr9\" (UniqueName: \"kubernetes.io/projected/85468a31-5db8-44d1-8027-afab2b6f8e02-kube-api-access-w2zr9\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545429 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-mountpoint-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545575 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-encryption-config\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545707 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9693007-9dd0-46b0-9f2d-b75421bfcde9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545806 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dadf4421-75bf-4cb8-8d63-040b446ecf0f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.545936 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfz4g\" (UniqueName: \"kubernetes.io/projected/45107232-8768-41e7-b274-4f17095bb302-kube-api-access-tfz4g\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546091 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546214 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgnp8\" (UniqueName: \"kubernetes.io/projected/c9693007-9dd0-46b0-9f2d-b75421bfcde9-kube-api-access-kgnp8\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546339 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-config\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546396 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6de60db-1a8c-4be4-9943-c5afcd8f4566-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546455 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eee929d7-38fc-464c-9ca1-a32f6e1771b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546493 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546526 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546556 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546610 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546643 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de60db-1a8c-4be4-9943-c5afcd8f4566-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546674 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-default-certificate\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546699 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-trusted-ca-bundle\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546745 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pdn6\" (UniqueName: \"kubernetes.io/projected/029287ae-ef37-46c9-b023-1d7074803fd6-kube-api-access-4pdn6\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546786 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd4lm\" (UniqueName: \"kubernetes.io/projected/a8a64945-22c6-496b-8e9c-fc1442e93554-kube-api-access-sd4lm\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546854 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-config-volume\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546909 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f695822d-7cf9-483b-bf22-ebc2bec117e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546949 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfl6m\" (UniqueName: \"kubernetes.io/projected/1f8f0912-b7a1-4e77-b270-4ba5e910cbf9-kube-api-access-lfl6m\") pod \"downloads-7954f5f757-ggnbw\" (UID: \"1f8f0912-b7a1-4e77-b270-4ba5e910cbf9\") " pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.546979 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547013 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf201413-6366-4074-8cc2-90819f77229c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547043 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-image-import-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547077 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdshq\" (UniqueName: \"kubernetes.io/projected/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-kube-api-access-pdshq\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547113 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/909b3993-ea89-4b5a-8376-4f17044468bb-serving-cert\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547165 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547392 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-stats-auth\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547442 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcvf2\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-kube-api-access-bcvf2\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547475 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-client\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547507 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d516372c-17a3-438c-87f8-d2e3edfb5670-proxy-tls\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547567 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzzz9\" (UniqueName: \"kubernetes.io/projected/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-kube-api-access-rzzz9\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547597 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547625 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547659 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6e26d7b-9ede-44a2-97d6-f60b80b38839-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547697 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547735 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-images\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547767 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xdkv\" (UniqueName: \"kubernetes.io/projected/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-kube-api-access-7xdkv\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547798 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547830 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-csi-data-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547863 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5059531-28eb-4fa0-a276-3d7310a814a6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547917 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-cabundle\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.547956 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d516372c-17a3-438c-87f8-d2e3edfb5670-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548118 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-service-ca\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548155 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-certs\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548191 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548226 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2c48\" (UniqueName: \"kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.548760 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.048710352 +0000 UTC m=+143.529013390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548805 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit-dir\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548841 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxv5f\" (UniqueName: \"kubernetes.io/projected/f695822d-7cf9-483b-bf22-ebc2bec117e8-kube-api-access-gxv5f\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548880 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbeac130-c061-4ae1-a609-8808eb849e38-metrics-tls\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548933 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548964 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjmwd\" (UniqueName: \"kubernetes.io/projected/16cdb795-c6e1-4469-adb6-2e839fc9a55f-kube-api-access-qjmwd\") pod \"migrator-59844c95c7-v6c8s\" (UID: \"16cdb795-c6e1-4469-adb6-2e839fc9a55f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.548996 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.549027 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8s6g\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-kube-api-access-t8s6g\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.549068 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-machine-approver-tls\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.549399 4800 request.go:700] Waited for 1.002618056s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-operator/secrets?fieldSelector=metadata.name%3Dmetrics-tls&limit=500&resourceVersion=0 Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.549944 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-service-ca\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.549096 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-trusted-ca\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550216 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68j6j\" (UniqueName: \"kubernetes.io/projected/909b3993-ea89-4b5a-8376-4f17044468bb-kube-api-access-68j6j\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550335 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550435 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-node-bootstrap-token\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550539 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550647 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgd85\" (UniqueName: \"kubernetes.io/projected/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-kube-api-access-rgd85\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550765 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-apiservice-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550935 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-srv-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.551373 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce65736c-5fd3-4d03-9cf2-e8759f560551-metrics-tls\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552146 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-socket-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552321 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhpjd\" (UniqueName: \"kubernetes.io/projected/dadf4421-75bf-4cb8-8d63-040b446ecf0f-kube-api-access-mhpjd\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552435 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-serving-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552546 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552654 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kswd\" (UniqueName: \"kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552751 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w7kx\" (UniqueName: \"kubernetes.io/projected/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-kube-api-access-7w7kx\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552845 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-auth-proxy-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.552981 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553087 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r44z7\" (UniqueName: \"kubernetes.io/projected/025d333e-13b9-49c2-b6b8-512fd070c4db-kube-api-access-r44z7\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553187 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b081f2a5-0dee-4ee9-b5af-564af14f3e69-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553279 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553370 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-bound-sa-token\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553482 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq4sw\" (UniqueName: \"kubernetes.io/projected/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-kube-api-access-zq4sw\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553669 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5e37c96-49b2-4450-898d-d01847863fff-service-ca-bundle\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.550264 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.551688 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.553851 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd89g\" (UniqueName: \"kubernetes.io/projected/5543f84b-8bd9-4464-9ee7-68ef8ec50985-kube-api-access-kd89g\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.554208 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.554515 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5mp5\" (UniqueName: \"kubernetes.io/projected/cf201413-6366-4074-8cc2-90819f77229c-kube-api-access-w5mp5\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.554793 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.555036 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eee929d7-38fc-464c-9ca1-a32f6e1771b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.555286 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-config\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.555557 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-metrics-tls\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.555732 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.555904 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-serving-cert\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556064 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9693007-9dd0-46b0-9f2d-b75421bfcde9-serving-cert\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556220 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2c4f\" (UniqueName: \"kubernetes.io/projected/e1735130-68e0-4d83-97e3-49be363ff8dd-kube-api-access-l2c4f\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556373 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f017341d-0ebc-442d-8980-5e2839c89c3a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556509 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-serving-cert\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556648 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-oauth-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.556800 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.557145 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.557359 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f89rt\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.557580 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.557862 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-policies\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558109 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-srv-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558262 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558019 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558433 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-proxy-tls\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558697 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.558731 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559040 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-config\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559229 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f5059531-28eb-4fa0-a276-3d7310a814a6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559415 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrxxn\" (UniqueName: \"kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559581 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559803 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-key\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.559999 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-webhook-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560142 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560285 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-metrics-certs\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560449 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560589 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-registration-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560739 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqrsd\" (UniqueName: \"kubernetes.io/projected/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-kube-api-access-rqrsd\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.560847 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561028 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-encryption-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561267 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhbbh\" (UniqueName: \"kubernetes.io/projected/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-kube-api-access-qhbbh\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561325 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-profile-collector-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561361 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eee929d7-38fc-464c-9ca1-a32f6e1771b5-config\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561392 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-cert\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561425 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f695822d-7cf9-483b-bf22-ebc2bec117e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561457 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561487 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561524 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/092d47e3-8067-4702-bbd2-1a1bdb429b67-tmpfs\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561558 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561676 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-oauth-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561719 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c86f998b-7292-4a18-a830-7d2c46dcdf2c-serving-cert\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.561815 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.562498 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.562749 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.563025 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.564327 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5543f84b-8bd9-4464-9ee7-68ef8ec50985-console-oauth-config\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.571565 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.592544 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.611661 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.632743 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.652707 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.663433 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.663807 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-srv-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.663867 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.163818154 +0000 UTC m=+143.644121232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.663995 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce65736c-5fd3-4d03-9cf2-e8759f560551-metrics-tls\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664073 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhpjd\" (UniqueName: \"kubernetes.io/projected/dadf4421-75bf-4cb8-8d63-040b446ecf0f-kube-api-access-mhpjd\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664120 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-socket-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664189 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-serving-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664248 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664300 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kswd\" (UniqueName: \"kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664352 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w7kx\" (UniqueName: \"kubernetes.io/projected/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-kube-api-access-7w7kx\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664407 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-auth-proxy-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664466 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664522 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r44z7\" (UniqueName: \"kubernetes.io/projected/025d333e-13b9-49c2-b6b8-512fd070c4db-kube-api-access-r44z7\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664561 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-socket-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664575 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b081f2a5-0dee-4ee9-b5af-564af14f3e69-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664732 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664800 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-bound-sa-token\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664818 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664858 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq4sw\" (UniqueName: \"kubernetes.io/projected/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-kube-api-access-zq4sw\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.664976 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5e37c96-49b2-4450-898d-d01847863fff-service-ca-bundle\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665075 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665134 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd89g\" (UniqueName: \"kubernetes.io/projected/5543f84b-8bd9-4464-9ee7-68ef8ec50985-kube-api-access-kd89g\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665191 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665246 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eee929d7-38fc-464c-9ca1-a32f6e1771b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665296 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-config\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665316 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-service-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665347 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-metrics-tls\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665406 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5mp5\" (UniqueName: \"kubernetes.io/projected/cf201413-6366-4074-8cc2-90819f77229c-kube-api-access-w5mp5\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665464 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-serving-cert\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665522 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9693007-9dd0-46b0-9f2d-b75421bfcde9-serving-cert\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665578 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2c4f\" (UniqueName: \"kubernetes.io/projected/e1735130-68e0-4d83-97e3-49be363ff8dd-kube-api-access-l2c4f\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665629 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f017341d-0ebc-442d-8980-5e2839c89c3a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665681 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-serving-cert\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665740 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665739 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-serving-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665770 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-auth-proxy-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.665790 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666164 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-oauth-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666194 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-policies\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666215 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-srv-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666256 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666276 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-proxy-tls\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666300 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-config\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666302 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5e37c96-49b2-4450-898d-d01847863fff-service-ca-bundle\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666334 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrxxn\" (UniqueName: \"kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666355 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666378 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f5059531-28eb-4fa0-a276-3d7310a814a6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666397 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-key\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666414 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-webhook-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666433 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666456 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-metrics-certs\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666482 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-registration-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666503 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqrsd\" (UniqueName: \"kubernetes.io/projected/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-kube-api-access-rqrsd\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666524 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-encryption-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666544 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666574 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhbbh\" (UniqueName: \"kubernetes.io/projected/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-kube-api-access-qhbbh\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666599 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-profile-collector-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666617 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eee929d7-38fc-464c-9ca1-a32f6e1771b5-config\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666637 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f695822d-7cf9-483b-bf22-ebc2bec117e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666652 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-cert\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666668 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666686 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666702 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/092d47e3-8067-4702-bbd2-1a1bdb429b67-tmpfs\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666722 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c86f998b-7292-4a18-a830-7d2c46dcdf2c-serving-cert\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666748 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce65736c-5fd3-4d03-9cf2-e8759f560551-trusted-ca\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666769 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czrzm\" (UniqueName: \"kubernetes.io/projected/b081f2a5-0dee-4ee9-b5af-564af14f3e69-kube-api-access-czrzm\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666793 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znntx\" (UniqueName: \"kubernetes.io/projected/d516372c-17a3-438c-87f8-d2e3edfb5670-kube-api-access-znntx\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666809 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666828 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hkrs\" (UniqueName: \"kubernetes.io/projected/c86f998b-7292-4a18-a830-7d2c46dcdf2c-kube-api-access-5hkrs\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666844 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-client\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666862 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-serving-cert\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666926 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/029287ae-ef37-46c9-b023-1d7074803fd6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666952 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6de60db-1a8c-4be4-9943-c5afcd8f4566-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666976 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v65qb\" (UniqueName: \"kubernetes.io/projected/f017341d-0ebc-442d-8980-5e2839c89c3a-kube-api-access-v65qb\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666998 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666992 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f017341d-0ebc-442d-8980-5e2839c89c3a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667020 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-node-pullsecrets\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667043 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f017341d-0ebc-442d-8980-5e2839c89c3a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667066 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e26d7b-9ede-44a2-97d6-f60b80b38839-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667087 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gkd\" (UniqueName: \"kubernetes.io/projected/a5e37c96-49b2-4450-898d-d01847863fff-kube-api-access-57gkd\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667118 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88jt8\" (UniqueName: \"kubernetes.io/projected/092d47e3-8067-4702-bbd2-1a1bdb429b67-kube-api-access-88jt8\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667145 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-plugins-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667164 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667184 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667206 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzwkz\" (UniqueName: \"kubernetes.io/projected/dbeac130-c061-4ae1-a609-8808eb849e38-kube-api-access-gzwkz\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667221 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e26d7b-9ede-44a2-97d6-f60b80b38839-config\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667237 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-dir\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667256 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-mountpoint-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667273 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-encryption-config\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667291 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2zr9\" (UniqueName: \"kubernetes.io/projected/85468a31-5db8-44d1-8027-afab2b6f8e02-kube-api-access-w2zr9\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667308 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9693007-9dd0-46b0-9f2d-b75421bfcde9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667326 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dadf4421-75bf-4cb8-8d63-040b446ecf0f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667346 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfz4g\" (UniqueName: \"kubernetes.io/projected/45107232-8768-41e7-b274-4f17095bb302-kube-api-access-tfz4g\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667366 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667381 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgnp8\" (UniqueName: \"kubernetes.io/projected/c9693007-9dd0-46b0-9f2d-b75421bfcde9-kube-api-access-kgnp8\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667398 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-config\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667416 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6de60db-1a8c-4be4-9943-c5afcd8f4566-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667435 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eee929d7-38fc-464c-9ca1-a32f6e1771b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667455 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667473 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667491 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667517 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de60db-1a8c-4be4-9943-c5afcd8f4566-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667534 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-default-certificate\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667554 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd4lm\" (UniqueName: \"kubernetes.io/projected/a8a64945-22c6-496b-8e9c-fc1442e93554-kube-api-access-sd4lm\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667572 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-config-volume\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667589 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f695822d-7cf9-483b-bf22-ebc2bec117e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667607 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pdn6\" (UniqueName: \"kubernetes.io/projected/029287ae-ef37-46c9-b023-1d7074803fd6-kube-api-access-4pdn6\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667627 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfl6m\" (UniqueName: \"kubernetes.io/projected/1f8f0912-b7a1-4e77-b270-4ba5e910cbf9-kube-api-access-lfl6m\") pod \"downloads-7954f5f757-ggnbw\" (UID: \"1f8f0912-b7a1-4e77-b270-4ba5e910cbf9\") " pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667644 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667664 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-image-import-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667684 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdshq\" (UniqueName: \"kubernetes.io/projected/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-kube-api-access-pdshq\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667701 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf201413-6366-4074-8cc2-90819f77229c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667730 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/909b3993-ea89-4b5a-8376-4f17044468bb-serving-cert\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667754 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-stats-auth\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667780 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcvf2\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-kube-api-access-bcvf2\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667774 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-policies\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667798 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-client\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.667819 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d516372c-17a3-438c-87f8-d2e3edfb5670-proxy-tls\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668036 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-config\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668166 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzzz9\" (UniqueName: \"kubernetes.io/projected/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-kube-api-access-rzzz9\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668232 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668291 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668350 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6e26d7b-9ede-44a2-97d6-f60b80b38839-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668403 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-images\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668459 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xdkv\" (UniqueName: \"kubernetes.io/projected/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-kube-api-access-7xdkv\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668503 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5543f84b-8bd9-4464-9ee7-68ef8ec50985-oauth-serving-cert\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668525 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668584 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668638 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-csi-data-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668655 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-mountpoint-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668691 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5059531-28eb-4fa0-a276-3d7310a814a6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668742 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-cabundle\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668796 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d516372c-17a3-438c-87f8-d2e3edfb5670-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668877 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-certs\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.668972 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669029 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2c48\" (UniqueName: \"kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669092 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxv5f\" (UniqueName: \"kubernetes.io/projected/f695822d-7cf9-483b-bf22-ebc2bec117e8-kube-api-access-gxv5f\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669146 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit-dir\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669196 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669205 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbeac130-c061-4ae1-a609-8808eb849e38-metrics-tls\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669265 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669306 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-srv-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669324 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjmwd\" (UniqueName: \"kubernetes.io/projected/16cdb795-c6e1-4469-adb6-2e839fc9a55f-kube-api-access-qjmwd\") pod \"migrator-59844c95c7-v6c8s\" (UID: \"16cdb795-c6e1-4469-adb6-2e839fc9a55f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669379 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669433 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8s6g\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-kube-api-access-t8s6g\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669527 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-trusted-ca\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669584 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68j6j\" (UniqueName: \"kubernetes.io/projected/909b3993-ea89-4b5a-8376-4f17044468bb-kube-api-access-68j6j\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669639 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669694 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-node-bootstrap-token\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669763 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669762 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-machine-approver-tls\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669870 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669951 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgd85\" (UniqueName: \"kubernetes.io/projected/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-kube-api-access-rgd85\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.669986 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-apiservice-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.670175 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de60db-1a8c-4be4-9943-c5afcd8f4566-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.670481 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.670833 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/092d47e3-8067-4702-bbd2-1a1bdb429b67-tmpfs\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.671222 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.671548 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit-dir\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.674016 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c9693007-9dd0-46b0-9f2d-b75421bfcde9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.674712 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.174680706 +0000 UTC m=+143.654983924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.676398 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.676502 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-registration-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.677420 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-machine-approver-tls\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.678132 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f5059531-28eb-4fa0-a276-3d7310a814a6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.678302 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d516372c-17a3-438c-87f8-d2e3edfb5670-proxy-tls\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.678695 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-config\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.679433 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/909b3993-ea89-4b5a-8376-4f17044468bb-serving-cert\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.679503 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce65736c-5fd3-4d03-9cf2-e8759f560551-trusted-ca\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.679524 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/909b3993-ea89-4b5a-8376-4f17044468bb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.679703 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.679992 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.680184 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-csi-data-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.680958 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-stats-auth\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.681225 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.681509 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbeac130-c061-4ae1-a609-8808eb849e38-metrics-tls\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.681540 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5059531-28eb-4fa0-a276-3d7310a814a6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.681717 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f695822d-7cf9-483b-bf22-ebc2bec117e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.682295 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eee929d7-38fc-464c-9ca1-a32f6e1771b5-config\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.682451 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.682534 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-serving-cert\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.682729 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85468a31-5db8-44d1-8027-afab2b6f8e02-audit-dir\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.666992 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.683098 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.683155 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e26d7b-9ede-44a2-97d6-f60b80b38839-config\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.683305 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-node-pullsecrets\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.683335 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/025d333e-13b9-49c2-b6b8-512fd070c4db-plugins-dir\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.685349 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce65736c-5fd3-4d03-9cf2-e8759f560551-metrics-tls\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.685579 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-default-certificate\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.685617 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-client\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.686353 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.686615 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-profile-collector-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.686836 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.687244 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6e26d7b-9ede-44a2-97d6-f60b80b38839-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.687590 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f695822d-7cf9-483b-bf22-ebc2bec117e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.687855 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-encryption-config\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.687926 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.687959 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eee929d7-38fc-464c-9ca1-a32f6e1771b5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.688118 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9693007-9dd0-46b0-9f2d-b75421bfcde9-serving-cert\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.688582 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c86f998b-7292-4a18-a830-7d2c46dcdf2c-serving-cert\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.688870 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/85468a31-5db8-44d1-8027-afab2b6f8e02-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.689262 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f017341d-0ebc-442d-8980-5e2839c89c3a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.689636 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-audit\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.689769 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/45107232-8768-41e7-b274-4f17095bb302-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.690216 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.690443 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.690810 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-image-import-ca\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.691274 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.691402 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85468a31-5db8-44d1-8027-afab2b6f8e02-serving-cert\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.691518 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-config\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.691548 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.691914 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.692687 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-encryption-config\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.692748 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.693667 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d516372c-17a3-438c-87f8-d2e3edfb5670-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.693738 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b081f2a5-0dee-4ee9-b5af-564af14f3e69-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.694191 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.694277 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c86f998b-7292-4a18-a830-7d2c46dcdf2c-trusted-ca\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.694997 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-etcd-client\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.696447 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dadf4421-75bf-4cb8-8d63-040b446ecf0f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.698696 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6de60db-1a8c-4be4-9943-c5afcd8f4566-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.699573 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5e37c96-49b2-4450-898d-d01847863fff-metrics-certs\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.713216 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.733489 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.752420 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.756928 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-images\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.770639 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.771479 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.771692 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.271662249 +0000 UTC m=+143.751965287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.772349 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.772877 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.27286229 +0000 UTC m=+143.753165318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.781952 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-key\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.792052 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.801528 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-signing-cabundle\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.811745 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.830807 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.840525 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-proxy-tls\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.851817 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.874351 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.874566 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.374546259 +0000 UTC m=+143.854849297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.875062 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.875552 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.375543326 +0000 UTC m=+143.855846364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.886164 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx9fq\" (UniqueName: \"kubernetes.io/projected/d8e3932f-8cc3-4173-ba81-ec05e09cc50d-kube-api-access-wx9fq\") pod \"etcd-operator-b45778765-tpn7g\" (UID: \"d8e3932f-8cc3-4173-ba81-ec05e09cc50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.907333 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wd2f\" (UniqueName: \"kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f\") pod \"image-pruner-29484000-kr6nl\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.926798 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbzg4\" (UniqueName: \"kubernetes.io/projected/72de5335-31e7-4e8a-8dee-2b5c8afdeab9-kube-api-access-nbzg4\") pod \"machine-api-operator-5694c8668f-hbblt\" (UID: \"72de5335-31e7-4e8a-8dee-2b5c8afdeab9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.942759 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcxq5\" (UniqueName: \"kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5\") pod \"route-controller-manager-6576b87f9c-88ls8\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.965692 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhj5\" (UniqueName: \"kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5\") pod \"controller-manager-879f6c89f-lxktv\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.967761 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.971203 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.974364 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.976094 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.976351 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.476314401 +0000 UTC m=+143.956617479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.977056 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:27 crc kubenswrapper[4800]: E0122 00:09:27.977580 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.477559724 +0000 UTC m=+143.957862802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:27 crc kubenswrapper[4800]: I0122 00:09:27.991361 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.000580 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e1735130-68e0-4d83-97e3-49be363ff8dd-srv-cert\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.012199 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.032064 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.041514 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-serving-cert\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.049367 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.051917 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.057118 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-config\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.075086 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.080075 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.080315 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.580276101 +0000 UTC m=+144.060579129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.080443 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.081273 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.581228507 +0000 UTC m=+144.061531585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.092220 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.113344 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.121409 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.128948 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.131172 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.138594 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf201413-6366-4074-8cc2-90819f77229c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.162741 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.166223 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.171543 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.181443 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.182062 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.182403 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.6823449 +0000 UTC m=+144.162647948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.182611 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.183094 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.683074609 +0000 UTC m=+144.163377647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.191152 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.205027 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hbblt"] Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.212116 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: W0122 00:09:28.223663 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72de5335_31e7_4e8a_8dee_2b5c8afdeab9.slice/crio-ad45ce80b902e4a3d6a2b2fdd0e652e2102e98b32cd637c471544fa8d6500933 WatchSource:0}: Error finding container ad45ce80b902e4a3d6a2b2fdd0e652e2102e98b32cd637c471544fa8d6500933: Status 404 returned error can't find the container with id ad45ce80b902e4a3d6a2b2fdd0e652e2102e98b32cd637c471544fa8d6500933 Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.231205 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.231695 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29484000-kr6nl"] Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.244040 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-webhook-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.245064 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/092d47e3-8067-4702-bbd2-1a1bdb429b67-apiservice-cert\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.251598 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.258411 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:28 crc kubenswrapper[4800]: W0122 00:09:28.265112 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6adcf93_e68c_4766_85f9_948c65a0d397.slice/crio-c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9 WatchSource:0}: Error finding container c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9: Status 404 returned error can't find the container with id c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9 Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.271622 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.276643 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.284024 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.284836 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.78481459 +0000 UTC m=+144.265117628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.296950 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.312816 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.313355 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/029287ae-ef37-46c9-b023-1d7074803fd6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.320855 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-metrics-tls\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.322667 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tpn7g"] Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.331967 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.352037 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 00:09:28 crc kubenswrapper[4800]: W0122 00:09:28.360411 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e3932f_8cc3_4173_ba81_ec05e09cc50d.slice/crio-e946143adadfdaf048dddc902a91d2bab74ba5e7084a2b2931c3c84616362da8 WatchSource:0}: Error finding container e946143adadfdaf048dddc902a91d2bab74ba5e7084a2b2931c3c84616362da8: Status 404 returned error can't find the container with id e946143adadfdaf048dddc902a91d2bab74ba5e7084a2b2931c3c84616362da8 Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.364271 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-config-volume\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.373214 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.385743 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.386305 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:28.886281364 +0000 UTC m=+144.366584402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.391148 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.391613 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-certs\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.399544 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.411209 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.415802 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a8a64945-22c6-496b-8e9c-fc1442e93554-node-bootstrap-token\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.431082 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.451062 4800 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.471502 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.490549 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.500124 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.501341 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.001312462 +0000 UTC m=+144.481615500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.512385 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.531728 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.544924 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-cert\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.549797 4800 request.go:700] Waited for 1.894626942s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.552382 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.603453 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.604208 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.104191512 +0000 UTC m=+144.584494560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.607251 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.626581 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f89rt\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.648609 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhpjd\" (UniqueName: \"kubernetes.io/projected/dadf4421-75bf-4cb8-8d63-040b446ecf0f-kube-api-access-mhpjd\") pod \"multus-admission-controller-857f4d67dd-wv42v\" (UID: \"dadf4421-75bf-4cb8-8d63-040b446ecf0f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.666661 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kswd\" (UniqueName: \"kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd\") pod \"marketplace-operator-79b997595-r9z5h\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.687458 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r44z7\" (UniqueName: \"kubernetes.io/projected/025d333e-13b9-49c2-b6b8-512fd070c4db-kube-api-access-r44z7\") pod \"csi-hostpathplugin-7gmzt\" (UID: \"025d333e-13b9-49c2-b6b8-512fd070c4db\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.704767 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.705033 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.204998868 +0000 UTC m=+144.685301906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.705378 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.705764 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.205755929 +0000 UTC m=+144.686058967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.706484 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w7kx\" (UniqueName: \"kubernetes.io/projected/8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e-kube-api-access-7w7kx\") pod \"service-ca-operator-777779d784-lpx9z\" (UID: \"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.715583 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" event={"ID":"d8e3932f-8cc3-4173-ba81-ec05e09cc50d","Type":"ContainerStarted","Data":"ace604f285429b15afba01d0af76580ea292525f9a23d1632e6c5fc0a0ba7836"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.716069 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" event={"ID":"d8e3932f-8cc3-4173-ba81-ec05e09cc50d","Type":"ContainerStarted","Data":"e946143adadfdaf048dddc902a91d2bab74ba5e7084a2b2931c3c84616362da8"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.717835 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29484000-kr6nl" event={"ID":"d6adcf93-e68c-4766-85f9-948c65a0d397","Type":"ContainerStarted","Data":"7ab6ee4a6dbb795fbc0e6fb0d94d0ee039251f7180fa495ab2a5cb11c82d6b19"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.717918 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29484000-kr6nl" event={"ID":"d6adcf93-e68c-4766-85f9-948c65a0d397","Type":"ContainerStarted","Data":"c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.720783 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" event={"ID":"72de5335-31e7-4e8a-8dee-2b5c8afdeab9","Type":"ContainerStarted","Data":"691f60cb723c9460d9017563e1ecf88ef701e9c1e35e4e066364a6f4e3b0074d"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.720845 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" event={"ID":"72de5335-31e7-4e8a-8dee-2b5c8afdeab9","Type":"ContainerStarted","Data":"76ccff6ece38f9851e7951fda4394acd265230f4066aada000b455f5d6bef726"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.720861 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" event={"ID":"72de5335-31e7-4e8a-8dee-2b5c8afdeab9","Type":"ContainerStarted","Data":"ad45ce80b902e4a3d6a2b2fdd0e652e2102e98b32cd637c471544fa8d6500933"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.722629 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" event={"ID":"9f5ddfd1-20c3-4773-a581-324aad7c3d3f","Type":"ContainerStarted","Data":"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.722806 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" event={"ID":"9f5ddfd1-20c3-4773-a581-324aad7c3d3f","Type":"ContainerStarted","Data":"0dcde9e62663fa1ac328caca41380e23ee362a5f1cc3f58e9f371f7df77d577b"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.723155 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.724184 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" event={"ID":"5bcffd47-0f44-4095-892e-c3cf4acd975a","Type":"ContainerStarted","Data":"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.724220 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" event={"ID":"5bcffd47-0f44-4095-892e-c3cf4acd975a","Type":"ContainerStarted","Data":"f827a6b35faa68019dfc8e802ebec4dafb40459e085d6dcd32eda29d282ac095"} Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.724511 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.725752 4800 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-88ls8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.725813 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.725995 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-bound-sa-token\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.726115 4800 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lxktv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.726177 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.747358 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq4sw\" (UniqueName: \"kubernetes.io/projected/0ea8fa19-f1cc-4bd0-a812-14513c0326ca-kube-api-access-zq4sw\") pod \"service-ca-9c57cc56f-42dwh\" (UID: \"0ea8fa19-f1cc-4bd0-a812-14513c0326ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.767588 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5mp5\" (UniqueName: \"kubernetes.io/projected/cf201413-6366-4074-8cc2-90819f77229c-kube-api-access-w5mp5\") pod \"control-plane-machine-set-operator-78cbb6b69f-6grhd\" (UID: \"cf201413-6366-4074-8cc2-90819f77229c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.786862 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.806867 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.807022 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.306996965 +0000 UTC m=+144.787300003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.807287 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.809275 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.309249956 +0000 UTC m=+144.789553004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.814825 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd89g\" (UniqueName: \"kubernetes.io/projected/5543f84b-8bd9-4464-9ee7-68ef8ec50985-kube-api-access-kd89g\") pod \"console-f9d7485db-q6r6s\" (UID: \"5543f84b-8bd9-4464-9ee7-68ef8ec50985\") " pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.829847 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2c4f\" (UniqueName: \"kubernetes.io/projected/e1735130-68e0-4d83-97e3-49be363ff8dd-kube-api-access-l2c4f\") pod \"catalog-operator-68c6474976-8rnqq\" (UID: \"e1735130-68e0-4d83-97e3-49be363ff8dd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.852918 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrxxn\" (UniqueName: \"kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn\") pod \"collect-profiles-29484000-vjj9f\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.859365 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.866559 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.871217 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgd85\" (UniqueName: \"kubernetes.io/projected/b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6-kube-api-access-rgd85\") pod \"machine-config-operator-74547568cd-5qb2b\" (UID: \"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.873107 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.885077 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.889555 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.890564 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcvf2\" (UniqueName: \"kubernetes.io/projected/f5059531-28eb-4fa0-a276-3d7310a814a6-kube-api-access-bcvf2\") pod \"cluster-image-registry-operator-dc59b4c8b-vdggf\" (UID: \"f5059531-28eb-4fa0-a276-3d7310a814a6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.896954 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.904371 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.904740 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjmwd\" (UniqueName: \"kubernetes.io/projected/16cdb795-c6e1-4469-adb6-2e839fc9a55f-kube-api-access-qjmwd\") pod \"migrator-59844c95c7-v6c8s\" (UID: \"16cdb795-c6e1-4469-adb6-2e839fc9a55f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.908657 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.909014 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.408984852 +0000 UTC m=+144.889288080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.909141 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:28 crc kubenswrapper[4800]: E0122 00:09:28.910167 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.410146674 +0000 UTC m=+144.890449712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.923444 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.927949 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2c48\" (UniqueName: \"kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48\") pod \"oauth-openshift-558db77b4-bq4b8\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.954151 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd4lm\" (UniqueName: \"kubernetes.io/projected/a8a64945-22c6-496b-8e9c-fc1442e93554-kube-api-access-sd4lm\") pod \"machine-config-server-4hl4v\" (UID: \"a8a64945-22c6-496b-8e9c-fc1442e93554\") " pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.954497 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4hl4v" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.975337 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxv5f\" (UniqueName: \"kubernetes.io/projected/f695822d-7cf9-483b-bf22-ebc2bec117e8-kube-api-access-gxv5f\") pod \"openshift-controller-manager-operator-756b6f6bc6-csf2g\" (UID: \"f695822d-7cf9-483b-bf22-ebc2bec117e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:28 crc kubenswrapper[4800]: I0122 00:09:28.986260 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.003420 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgnp8\" (UniqueName: \"kubernetes.io/projected/c9693007-9dd0-46b0-9f2d-b75421bfcde9-kube-api-access-kgnp8\") pod \"openshift-config-operator-7777fb866f-x99cx\" (UID: \"c9693007-9dd0-46b0-9f2d-b75421bfcde9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.018957 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.019494 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.519473278 +0000 UTC m=+144.999776316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.028124 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2zr9\" (UniqueName: \"kubernetes.io/projected/85468a31-5db8-44d1-8027-afab2b6f8e02-kube-api-access-w2zr9\") pod \"apiserver-7bbb656c7d-fmb8f\" (UID: \"85468a31-5db8-44d1-8027-afab2b6f8e02\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.029140 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.044202 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.047118 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.060874 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.070022 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eee929d7-38fc-464c-9ca1-a32f6e1771b5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sjdlp\" (UID: \"eee929d7-38fc-464c-9ca1-a32f6e1771b5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.070453 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfz4g\" (UniqueName: \"kubernetes.io/projected/45107232-8768-41e7-b274-4f17095bb302-kube-api-access-tfz4g\") pod \"olm-operator-6b444d44fb-n9cq8\" (UID: \"45107232-8768-41e7-b274-4f17095bb302\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.080441 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pdn6\" (UniqueName: \"kubernetes.io/projected/029287ae-ef37-46c9-b023-1d7074803fd6-kube-api-access-4pdn6\") pod \"package-server-manager-789f6589d5-rlm9b\" (UID: \"029287ae-ef37-46c9-b023-1d7074803fd6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.093310 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.099213 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.101204 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzzz9\" (UniqueName: \"kubernetes.io/projected/a3b9dd2e-dc40-4468-98a5-f52cd6d7c737-kube-api-access-rzzz9\") pod \"machine-approver-56656f9798-4kd6f\" (UID: \"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.114230 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.120430 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfl6m\" (UniqueName: \"kubernetes.io/projected/1f8f0912-b7a1-4e77-b270-4ba5e910cbf9-kube-api-access-lfl6m\") pod \"downloads-7954f5f757-ggnbw\" (UID: \"1f8f0912-b7a1-4e77-b270-4ba5e910cbf9\") " pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.121483 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.121836 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.621823045 +0000 UTC m=+145.102126083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.135334 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.135758 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czrzm\" (UniqueName: \"kubernetes.io/projected/b081f2a5-0dee-4ee9-b5af-564af14f3e69-kube-api-access-czrzm\") pod \"cluster-samples-operator-665b6dd947-lvwxh\" (UID: \"b081f2a5-0dee-4ee9-b5af-564af14f3e69\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.154194 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.162936 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znntx\" (UniqueName: \"kubernetes.io/projected/d516372c-17a3-438c-87f8-d2e3edfb5670-kube-api-access-znntx\") pod \"machine-config-controller-84d6567774-shdlf\" (UID: \"d516372c-17a3-438c-87f8-d2e3edfb5670\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.183709 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqrsd\" (UniqueName: \"kubernetes.io/projected/6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7-kube-api-access-rqrsd\") pod \"apiserver-76f77b778f-zn8zh\" (UID: \"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7\") " pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.209682 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hkrs\" (UniqueName: \"kubernetes.io/projected/c86f998b-7292-4a18-a830-7d2c46dcdf2c-kube-api-access-5hkrs\") pod \"console-operator-58897d9998-6t2gz\" (UID: \"c86f998b-7292-4a18-a830-7d2c46dcdf2c\") " pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.222183 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.223496 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.723463062 +0000 UTC m=+145.203766100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.234151 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.234681 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6e26d7b-9ede-44a2-97d6-f60b80b38839-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zsr9k\" (UID: \"e6e26d7b-9ede-44a2-97d6-f60b80b38839\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.250373 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.258424 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.259314 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xdkv\" (UniqueName: \"kubernetes.io/projected/25c87458-5680-441e-8d4f-8f8b3d0ea5d5-kube-api-access-7xdkv\") pod \"dns-default-6l6mb\" (UID: \"25c87458-5680-441e-8d4f-8f8b3d0ea5d5\") " pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.259594 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8s6g\" (UniqueName: \"kubernetes.io/projected/ce65736c-5fd3-4d03-9cf2-e8759f560551-kube-api-access-t8s6g\") pod \"ingress-operator-5b745b69d9-v5n9s\" (UID: \"ce65736c-5fd3-4d03-9cf2-e8759f560551\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.277075 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhbbh\" (UniqueName: \"kubernetes.io/projected/7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0-kube-api-access-qhbbh\") pod \"ingress-canary-g5tbw\" (UID: \"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0\") " pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.292880 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g5tbw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.316930 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68j6j\" (UniqueName: \"kubernetes.io/projected/909b3993-ea89-4b5a-8376-4f17044468bb-kube-api-access-68j6j\") pod \"authentication-operator-69f744f599-shzt2\" (UID: \"909b3993-ea89-4b5a-8376-4f17044468bb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.323607 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.325454 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.325807 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.8257948 +0000 UTC m=+145.306097838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.341630 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdshq\" (UniqueName: \"kubernetes.io/projected/29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8-kube-api-access-pdshq\") pod \"kube-storage-version-migrator-operator-b67b599dd-6s75r\" (UID: \"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.353858 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.373942 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v65qb\" (UniqueName: \"kubernetes.io/projected/f017341d-0ebc-442d-8980-5e2839c89c3a-kube-api-access-v65qb\") pod \"openshift-apiserver-operator-796bbdcf4f-nzbgw\" (UID: \"f017341d-0ebc-442d-8980-5e2839c89c3a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.374425 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6de60db-1a8c-4be4-9943-c5afcd8f4566-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-msgt9\" (UID: \"e6de60db-1a8c-4be4-9943-c5afcd8f4566\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.385247 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.385524 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88jt8\" (UniqueName: \"kubernetes.io/projected/092d47e3-8067-4702-bbd2-1a1bdb429b67-kube-api-access-88jt8\") pod \"packageserver-d55dfcdfc-fzt4x\" (UID: \"092d47e3-8067-4702-bbd2-1a1bdb429b67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.401479 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.402783 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzwkz\" (UniqueName: \"kubernetes.io/projected/dbeac130-c061-4ae1-a609-8808eb849e38-kube-api-access-gzwkz\") pod \"dns-operator-744455d44c-cv9lb\" (UID: \"dbeac130-c061-4ae1-a609-8808eb849e38\") " pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.405164 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.411024 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wv42v"] Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.430964 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.431927 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.432327 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:29.932294188 +0000 UTC m=+145.412597226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.443466 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gkd\" (UniqueName: \"kubernetes.io/projected/a5e37c96-49b2-4450-898d-d01847863fff-kube-api-access-57gkd\") pod \"router-default-5444994796-ffvfb\" (UID: \"a5e37c96-49b2-4450-898d-d01847863fff\") " pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.443742 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.494606 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.502301 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.521404 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.522723 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.551249 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.552597 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.553100 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.053085331 +0000 UTC m=+145.533388369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.654013 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.655157 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.15511642 +0000 UTC m=+145.635419458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.675941 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.677348 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.680962 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b"] Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.723398 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.745457 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4hl4v" event={"ID":"a8a64945-22c6-496b-8e9c-fc1442e93554","Type":"ContainerStarted","Data":"5aa1a3199cbebe8726aa213863682dbf6707bbdf53033b96b811f43acc0ee5ec"} Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.745507 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4hl4v" event={"ID":"a8a64945-22c6-496b-8e9c-fc1442e93554","Type":"ContainerStarted","Data":"7efd413ebf7611621dd87a92e430cc57a9febbc56cb6de05e2e088e2fa784fed"} Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.754755 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.754916 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.755505 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.755836 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.255821942 +0000 UTC m=+145.736124980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.857188 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.857400 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.357367097 +0000 UTC m=+145.837670135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.857724 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.859086 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.359076613 +0000 UTC m=+145.839379861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:29 crc kubenswrapper[4800]: I0122 00:09:29.958902 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:29 crc kubenswrapper[4800]: E0122 00:09:29.959321 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.459303763 +0000 UTC m=+145.939606801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.061251 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.061714 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.561697931 +0000 UTC m=+146.042000969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.165521 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.166049 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.666026081 +0000 UTC m=+146.146329119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.275682 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.276192 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.776164757 +0000 UTC m=+146.256467795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.315426 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.334298 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.340224 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.359003 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.380335 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-42dwh"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.380391 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.380478 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.880456477 +0000 UTC m=+146.360759515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.393995 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.394992 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.894960497 +0000 UTC m=+146.375263535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.405089 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f"] Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.497629 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.497837 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.997803088 +0000 UTC m=+146.478106126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.498743 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.499216 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:30.999206776 +0000 UTC m=+146.479509814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.599809 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.600348 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.100328109 +0000 UTC m=+146.580631147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.704873 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.705660 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.205644896 +0000 UTC m=+146.685948044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.806872 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.807490 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.307453578 +0000 UTC m=+146.787756616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.818104 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" podStartSLOduration=128.818078524 podStartE2EDuration="2m8.818078524s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:30.774083419 +0000 UTC m=+146.254386457" watchObservedRunningTime="2026-01-22 00:09:30.818078524 +0000 UTC m=+146.298381552" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.908075 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.908115 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.908147 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.908177 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.908206 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:30 crc kubenswrapper[4800]: E0122 00:09:30.914585 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.414560513 +0000 UTC m=+146.894863551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.916562 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.964046 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.964651 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:30 crc kubenswrapper[4800]: I0122 00:09:30.965969 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.006872 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" podStartSLOduration=128.006853709 podStartE2EDuration="2m8.006853709s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:30.982192415 +0000 UTC m=+146.462495453" watchObservedRunningTime="2026-01-22 00:09:31.006853709 +0000 UTC m=+146.487156737" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.009516 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.010077 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.510054765 +0000 UTC m=+146.990357793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.048234 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.079251 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.084458 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hbblt" podStartSLOduration=128.084435889 podStartE2EDuration="2m8.084435889s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.048217413 +0000 UTC m=+146.528520451" watchObservedRunningTime="2026-01-22 00:09:31.084435889 +0000 UTC m=+146.564738927" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.090934 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.104933 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ffvfb" event={"ID":"a5e37c96-49b2-4450-898d-d01847863fff","Type":"ContainerStarted","Data":"d615c20497f491ef1be55c6a4731f51a762f71b3dcd7cf7538443e813acf6e85"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105184 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ffvfb" event={"ID":"a5e37c96-49b2-4450-898d-d01847863fff","Type":"ContainerStarted","Data":"e44d090f7b476cb685ab5a7ebab0765c2fd4aee9ae7713c0d373d2d931fd9588"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105286 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" event={"ID":"dadf4421-75bf-4cb8-8d63-040b446ecf0f","Type":"ContainerStarted","Data":"37b90b0b3172a4812aca325c92f074703cc0dbd5a8dfc78ed7603e5be9ae457b"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105360 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" event={"ID":"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6","Type":"ContainerStarted","Data":"4225814f12dd4c96730597363b1494656f83d41796e64cd1f392e6be98b38d15"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105422 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" event={"ID":"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737","Type":"ContainerStarted","Data":"d5c117df7f101af3f665e7919eb5e25f0f13777ef9c694a14426e46a9b347c26"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105491 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" event={"ID":"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737","Type":"ContainerStarted","Data":"e279b73d4c7b50342f57b022bd78e930b6351431c0ee7fac9de2655280ce15b5"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.105555 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" event={"ID":"e1735130-68e0-4d83-97e3-49be363ff8dd","Type":"ContainerStarted","Data":"66e730e5aea2ca9930887e544e17a95a4674b4520a0c50bb610d7f1cb6322e61"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.115095 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.115566 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.615547457 +0000 UTC m=+147.095850495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.159597 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29484000-kr6nl" podStartSLOduration=129.159550862 podStartE2EDuration="2m9.159550862s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.084757578 +0000 UTC m=+146.565060616" watchObservedRunningTime="2026-01-22 00:09:31.159550862 +0000 UTC m=+146.639853920" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.214714 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4hl4v" podStartSLOduration=5.214688148 podStartE2EDuration="5.214688148s" podCreationTimestamp="2026-01-22 00:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.211612784 +0000 UTC m=+146.691915822" watchObservedRunningTime="2026-01-22 00:09:31.214688148 +0000 UTC m=+146.694991186" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.224325 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.225972 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.725948931 +0000 UTC m=+147.206251969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.309242 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-tpn7g" podStartSLOduration=128.309214413 podStartE2EDuration="2m8.309214413s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.307027775 +0000 UTC m=+146.787330813" watchObservedRunningTime="2026-01-22 00:09:31.309214413 +0000 UTC m=+146.789517451" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.332348 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.332822 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.832801709 +0000 UTC m=+147.313104747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.433408 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.434187 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:31.934167709 +0000 UTC m=+147.414470747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.535518 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.535870 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.035856598 +0000 UTC m=+147.516159636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.636718 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.636970 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.13690769 +0000 UTC m=+147.617210728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.637176 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.637584 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.137576678 +0000 UTC m=+147.617879716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.706144 4800 csr.go:261] certificate signing request csr-gf6v8 is approved, waiting to be issued Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.712475 4800 csr.go:257] certificate signing request csr-gf6v8 is issued Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.723697 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ffvfb" podStartSLOduration=128.723670827 podStartE2EDuration="2m8.723670827s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.382170418 +0000 UTC m=+146.862473466" watchObservedRunningTime="2026-01-22 00:09:31.723670827 +0000 UTC m=+147.203973865" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.726179 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.727577 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf"] Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.737988 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.738491 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.238470376 +0000 UTC m=+147.718773424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.753865 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmzt"] Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.784814 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q6r6s"] Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.839507 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.839869 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.339850996 +0000 UTC m=+147.820154034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.856515 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f"] Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.861781 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq4b8"] Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.915523 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" event={"ID":"dadf4421-75bf-4cb8-8d63-040b446ecf0f","Type":"ContainerStarted","Data":"8e7461d6bf71e06f0a6bfb126c27fc28a01fa030b86981cb6594f6fb054914d3"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.915569 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" event={"ID":"dadf4421-75bf-4cb8-8d63-040b446ecf0f","Type":"ContainerStarted","Data":"e4294f36fb27e557d7fba1c800770fdf4012ac99f9b17a555d9451a01694d19e"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.919200 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" event={"ID":"0f5bfc30-ad0a-486a-baad-9dc8249f7170","Type":"ContainerStarted","Data":"75627e11053cb23ff4dacefd3832ce37a3419f699288840153f5a35807fe7040"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.919274 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" event={"ID":"0f5bfc30-ad0a-486a-baad-9dc8249f7170","Type":"ContainerStarted","Data":"18496040af9a2f886482d376fb89e2d1dfeb519b8c57a6757a03fefbbc44540a"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.921359 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" event={"ID":"a3b9dd2e-dc40-4468-98a5-f52cd6d7c737","Type":"ContainerStarted","Data":"ad2086daa1700f808e34cd4aea7651034678a50e2d18740efc8ca53ee60d0435"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.930256 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:31 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:31 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:31 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.930299 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.938407 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" event={"ID":"7446e4e9-c6b6-48e5-8488-6484f9f274ae","Type":"ContainerStarted","Data":"ae93bae088e74583820f91b8ff08f1b07e5567e6a65e46b7ef4c836c367bc392"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.938462 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" event={"ID":"7446e4e9-c6b6-48e5-8488-6484f9f274ae","Type":"ContainerStarted","Data":"c20780e188eb3932d13b109f57b220106a94d787c436caa3bce067105cf66ed7"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.939013 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.940014 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:31 crc kubenswrapper[4800]: E0122 00:09:31.940308 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.440294952 +0000 UTC m=+147.920597990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.943723 4800 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r9z5h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.943779 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.943716 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-wv42v" podStartSLOduration=128.943692693 podStartE2EDuration="2m8.943692693s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.937851026 +0000 UTC m=+147.418154064" watchObservedRunningTime="2026-01-22 00:09:31.943692693 +0000 UTC m=+147.423995731" Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.947036 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" event={"ID":"0ea8fa19-f1cc-4bd0-a812-14513c0326ca","Type":"ContainerStarted","Data":"d8531ed1e36873573155717bac57c9be5c4763c759c4478cb23341485bac3e59"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.947132 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" event={"ID":"0ea8fa19-f1cc-4bd0-a812-14513c0326ca","Type":"ContainerStarted","Data":"ff94926a1de150b2e3f039466dc1c9eddc7172d6aa535e9506db3f05997121b7"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.978488 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" event={"ID":"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e","Type":"ContainerStarted","Data":"0c4d8c289cfa2dcaa00e76c63515231bbe5b3fdc3e3568ee25029fdac72b3817"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.978847 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" event={"ID":"8fd8dc7e-c0a4-4b98-b0ec-9948bcdab22e","Type":"ContainerStarted","Data":"85b4e97ba70f2f7deb63ef814b53419668661abe2dd13869270f0aa15cc322d2"} Jan 22 00:09:31 crc kubenswrapper[4800]: I0122 00:09:31.988122 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" podStartSLOduration=128.988101629 podStartE2EDuration="2m8.988101629s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:31.978729496 +0000 UTC m=+147.459032534" watchObservedRunningTime="2026-01-22 00:09:31.988101629 +0000 UTC m=+147.468404667" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.006017 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" event={"ID":"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6","Type":"ContainerStarted","Data":"18318f7a58041c9257ba3ce37794dc86f17481f6480530cfce58ad67033e6bd4"} Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.006062 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" event={"ID":"b3ad2a97-c0ad-4650-ad0c-cab0170bc1d6","Type":"ContainerStarted","Data":"9f07356e9f4e1df8d6e3fb41067ba09f9abb05c3fdad8a89bb4081c870c25a2c"} Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.009439 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4kd6f" podStartSLOduration=130.009422233 podStartE2EDuration="2m10.009422233s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.007682037 +0000 UTC m=+147.487985075" watchObservedRunningTime="2026-01-22 00:09:32.009422233 +0000 UTC m=+147.489725271" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.011968 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" event={"ID":"cf201413-6366-4074-8cc2-90819f77229c","Type":"ContainerStarted","Data":"b905a2f923116a95072e02eaabbc77b920b904c10040bed5ff7578b6c13b250b"} Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.012000 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" event={"ID":"cf201413-6366-4074-8cc2-90819f77229c","Type":"ContainerStarted","Data":"85a2d1950c83716cc3a9d7de3b2804c744edde9ab0d34cb314af52754184f0df"} Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.041611 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" event={"ID":"e1735130-68e0-4d83-97e3-49be363ff8dd","Type":"ContainerStarted","Data":"b00c9bb523e9e9507edad7d4c66915fc88121d71deb6d0c755382133b327fbf6"} Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.041689 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.042969 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.046711 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.546692528 +0000 UTC m=+148.026995566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.049716 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-42dwh" podStartSLOduration=129.049694698 podStartE2EDuration="2m9.049694698s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.048106796 +0000 UTC m=+147.528409834" watchObservedRunningTime="2026-01-22 00:09:32.049694698 +0000 UTC m=+147.529997736" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.050274 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" podStartSLOduration=129.050266813 podStartE2EDuration="2m9.050266813s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.027326866 +0000 UTC m=+147.507629904" watchObservedRunningTime="2026-01-22 00:09:32.050266813 +0000 UTC m=+147.530569851" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.073584 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.074991 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5qb2b" podStartSLOduration=129.074964889 podStartE2EDuration="2m9.074964889s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.07130058 +0000 UTC m=+147.551603618" watchObservedRunningTime="2026-01-22 00:09:32.074964889 +0000 UTC m=+147.555267927" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.103073 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpx9z" podStartSLOduration=129.103045305 podStartE2EDuration="2m9.103045305s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.094189137 +0000 UTC m=+147.574492165" watchObservedRunningTime="2026-01-22 00:09:32.103045305 +0000 UTC m=+147.583348343" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.128932 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8rnqq" podStartSLOduration=129.128914763 podStartE2EDuration="2m9.128914763s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.126280871 +0000 UTC m=+147.606583909" watchObservedRunningTime="2026-01-22 00:09:32.128914763 +0000 UTC m=+147.609217801" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.144559 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.145109 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.645083498 +0000 UTC m=+148.125386536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.151356 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6grhd" podStartSLOduration=129.151339527 podStartE2EDuration="2m9.151339527s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:32.15111282 +0000 UTC m=+147.631415858" watchObservedRunningTime="2026-01-22 00:09:32.151339527 +0000 UTC m=+147.631642555" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.177397 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.182204 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.200684 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b"] Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.211007 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16cdb795_c6e1_4469_adb6_2e839fc9a55f.slice/crio-0a80949be6371e4500db76b38893d0f4d517f2922f369dd8256d791f865593a8 WatchSource:0}: Error finding container 0a80949be6371e4500db76b38893d0f4d517f2922f369dd8256d791f865593a8: Status 404 returned error can't find the container with id 0a80949be6371e4500db76b38893d0f4d517f2922f369dd8256d791f865593a8 Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.215475 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod029287ae_ef37_46c9_b023_1d7074803fd6.slice/crio-2ff8b03dea3fb857d07a028032736775fd776e2d9e823a7b054923a7111e9710 WatchSource:0}: Error finding container 2ff8b03dea3fb857d07a028032736775fd776e2d9e823a7b054923a7111e9710: Status 404 returned error can't find the container with id 2ff8b03dea3fb857d07a028032736775fd776e2d9e823a7b054923a7111e9710 Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.246932 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.248415 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.748396 +0000 UTC m=+148.228699038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.349098 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.349809 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.849783351 +0000 UTC m=+148.330086389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.453228 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.454006 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:32.953988708 +0000 UTC m=+148.434291746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.505439 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ggnbw"] Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.510584 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f8f0912_b7a1_4e77_b270_4ba5e910cbf9.slice/crio-f73388f84dd8842fd7fe7a70048b413da9cf7dd208884e355f09ad8530f1d345 WatchSource:0}: Error finding container f73388f84dd8842fd7fe7a70048b413da9cf7dd208884e355f09ad8530f1d345: Status 404 returned error can't find the container with id f73388f84dd8842fd7fe7a70048b413da9cf7dd208884e355f09ad8530f1d345 Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.518788 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.533859 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.555410 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.555442 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g5tbw"] Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.555728 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.055699357 +0000 UTC m=+148.536002395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.564704 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-shzt2"] Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.565763 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45107232_8768_41e7_b274_4f17095bb302.slice/crio-371b6279a642ac29244e2038b83ee3f85812454dc834db8c521014f2deb89b5b WatchSource:0}: Error finding container 371b6279a642ac29244e2038b83ee3f85812454dc834db8c521014f2deb89b5b: Status 404 returned error can't find the container with id 371b6279a642ac29244e2038b83ee3f85812454dc834db8c521014f2deb89b5b Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.566521 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.583696 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x99cx"] Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.583830 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod909b3993_ea89_4b5a_8376_4f17044468bb.slice/crio-741acf24412dca4bbeea514b600bba3701cb1a7c25fad447c87ca29df001eda8 WatchSource:0}: Error finding container 741acf24412dca4bbeea514b600bba3701cb1a7c25fad447c87ca29df001eda8: Status 404 returned error can't find the container with id 741acf24412dca4bbeea514b600bba3701cb1a7c25fad447c87ca29df001eda8 Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.616658 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.661520 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.661866 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.161854817 +0000 UTC m=+148.642157855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.710277 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cv9lb"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.715328 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-22 00:04:31 +0000 UTC, rotation deadline is 2026-11-17 15:24:28.817712411 +0000 UTC Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.715363 4800 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7191h14m56.102352173s for next certificate rotation Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.739320 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd516372c_17a3_438c_87f8_d2e3edfb5670.slice/crio-7e477237dce3d8d362b0923d1f81194c8735b1bd4101505b0bed148223a4d64f WatchSource:0}: Error finding container 7e477237dce3d8d362b0923d1f81194c8735b1bd4101505b0bed148223a4d64f: Status 404 returned error can't find the container with id 7e477237dce3d8d362b0923d1f81194c8735b1bd4101505b0bed148223a4d64f Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.739629 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:32 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:32 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:32 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.739661 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.746963 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.755613 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.765139 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.765471 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.265453847 +0000 UTC m=+148.745756885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.782970 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.789477 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6l6mb"] Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.794534 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf017341d_0ebc_442d_8980_5e2839c89c3a.slice/crio-0adeda191359ed02fcee8e1fb615bf90dbe4f1e2176020edb9f41fb1bff4d71c WatchSource:0}: Error finding container 0adeda191359ed02fcee8e1fb615bf90dbe4f1e2176020edb9f41fb1bff4d71c: Status 404 returned error can't find the container with id 0adeda191359ed02fcee8e1fb615bf90dbe4f1e2176020edb9f41fb1bff4d71c Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.814662 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.867288 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.868288 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.368272717 +0000 UTC m=+148.848575755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.877993 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.878050 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zn8zh"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.878067 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6t2gz"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.923100 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s"] Jan 22 00:09:32 crc kubenswrapper[4800]: I0122 00:09:32.967990 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:32 crc kubenswrapper[4800]: E0122 00:09:32.968363 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.468346302 +0000 UTC m=+148.948649340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.995270 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-6674e8e54f18d6abccfcee50119b6016506935f6b434a6b3614413a6322ffdc9 WatchSource:0}: Error finding container 6674e8e54f18d6abccfcee50119b6016506935f6b434a6b3614413a6322ffdc9: Status 404 returned error can't find the container with id 6674e8e54f18d6abccfcee50119b6016506935f6b434a6b3614413a6322ffdc9 Jan 22 00:09:32 crc kubenswrapper[4800]: W0122 00:09:32.999625 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-998d8dfced1ea9982203f66381740704ef6dc6b51d772ad6a4ebf18552424b05 WatchSource:0}: Error finding container 998d8dfced1ea9982203f66381740704ef6dc6b51d772ad6a4ebf18552424b05: Status 404 returned error can't find the container with id 998d8dfced1ea9982203f66381740704ef6dc6b51d772ad6a4ebf18552424b05 Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.076262 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.076772 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.576756912 +0000 UTC m=+149.057059950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.092214 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" event={"ID":"e6de60db-1a8c-4be4-9943-c5afcd8f4566","Type":"ContainerStarted","Data":"1af17d3670a9cb42ccaf7a0207087e2d8280486e1858ff05c8cfde50d3a33128"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.092923 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" event={"ID":"ce65736c-5fd3-4d03-9cf2-e8759f560551","Type":"ContainerStarted","Data":"506217d2366a4121e219d7ec35866906cf1af02e81a15f0258b17905c66188bf"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.093519 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6l6mb" event={"ID":"25c87458-5680-441e-8d4f-8f8b3d0ea5d5","Type":"ContainerStarted","Data":"aeedf6c625a2a94690e50c5982cba3541b67c7b5d26a427276b469fc848fa3ab"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.094282 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"456605d09650bfa87ae81d236056d3aa0c13af0fbc7cade9ab3b02275b209cea"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.095386 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" event={"ID":"029287ae-ef37-46c9-b023-1d7074803fd6","Type":"ContainerStarted","Data":"8256b90b58efc890fffe760d2741eca94730083d76a9bdb2561e6e40b77b5240"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.095414 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" event={"ID":"029287ae-ef37-46c9-b023-1d7074803fd6","Type":"ContainerStarted","Data":"2ff8b03dea3fb857d07a028032736775fd776e2d9e823a7b054923a7111e9710"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.097839 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" event={"ID":"eee929d7-38fc-464c-9ca1-a32f6e1771b5","Type":"ContainerStarted","Data":"8664a27c57aa7d3602b434d62eb8ced3ba278eb519e47da7027192249285315f"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.116538 4800 generic.go:334] "Generic (PLEG): container finished" podID="85468a31-5db8-44d1-8027-afab2b6f8e02" containerID="0ae3c5642223cd3706c1933b46acd1cbe9cc99f35bc8d3e5e0910c2e783b5e52" exitCode=0 Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.116611 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" event={"ID":"85468a31-5db8-44d1-8027-afab2b6f8e02","Type":"ContainerDied","Data":"0ae3c5642223cd3706c1933b46acd1cbe9cc99f35bc8d3e5e0910c2e783b5e52"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.116641 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" event={"ID":"85468a31-5db8-44d1-8027-afab2b6f8e02","Type":"ContainerStarted","Data":"2d67ed496f3d32ffa23d5f3c76d401fc48f71b253b78d1cb12d747402cb1c15e"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.120526 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q6r6s" event={"ID":"5543f84b-8bd9-4464-9ee7-68ef8ec50985","Type":"ContainerStarted","Data":"16369a425e03d48e19a3b449962894a11a856fde822c0f6d9d03a6c84f5277ca"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.120577 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q6r6s" event={"ID":"5543f84b-8bd9-4464-9ee7-68ef8ec50985","Type":"ContainerStarted","Data":"ba8de51b899346346ce359f94c68131ea4c26c1f7ac43a2c3471064491f0e380"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.134512 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" event={"ID":"16cdb795-c6e1-4469-adb6-2e839fc9a55f","Type":"ContainerStarted","Data":"6860a33abfac2b2b6f4a99af58fcc8810de56b2d3883c30e3e3bd44b759ecc24"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.134562 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" event={"ID":"16cdb795-c6e1-4469-adb6-2e839fc9a55f","Type":"ContainerStarted","Data":"d77d43c001622cda82f047662074903edb976903873032474ac7c62faa2933be"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.134572 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" event={"ID":"16cdb795-c6e1-4469-adb6-2e839fc9a55f","Type":"ContainerStarted","Data":"0a80949be6371e4500db76b38893d0f4d517f2922f369dd8256d791f865593a8"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.154374 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" event={"ID":"c9693007-9dd0-46b0-9f2d-b75421bfcde9","Type":"ContainerStarted","Data":"0b9519b14e3ff3fc447641bfc600c8320d501180cdbd383322df5052c3b7e639"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.160099 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" event={"ID":"dbeac130-c061-4ae1-a609-8808eb849e38","Type":"ContainerStarted","Data":"fe0ad2b739c732b1351de236fb3b8e111fc025ecdd16a46a6b75b62c27bd421b"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.163869 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"998d8dfced1ea9982203f66381740704ef6dc6b51d772ad6a4ebf18552424b05"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.174830 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" event={"ID":"45107232-8768-41e7-b274-4f17095bb302","Type":"ContainerStarted","Data":"fbd31100b8c0ae9e4997cfefb6588155fa0abb04dbad5b6524a37df10a4a1b61"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.174897 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" event={"ID":"45107232-8768-41e7-b274-4f17095bb302","Type":"ContainerStarted","Data":"371b6279a642ac29244e2038b83ee3f85812454dc834db8c521014f2deb89b5b"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.175855 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.177059 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.177210 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.677175107 +0000 UTC m=+149.157478145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.177261 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.177564 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.677550997 +0000 UTC m=+149.157854035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.178540 4800 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-n9cq8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.178573 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" podUID="45107232-8768-41e7-b274-4f17095bb302" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.185645 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-q6r6s" podStartSLOduration=131.185631545 podStartE2EDuration="2m11.185631545s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.185134152 +0000 UTC m=+148.665437190" watchObservedRunningTime="2026-01-22 00:09:33.185631545 +0000 UTC m=+148.665934583" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.190211 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" event={"ID":"909b3993-ea89-4b5a-8376-4f17044468bb","Type":"ContainerStarted","Data":"741acf24412dca4bbeea514b600bba3701cb1a7c25fad447c87ca29df001eda8"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.202643 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" event={"ID":"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7","Type":"ContainerStarted","Data":"9a3d66d133e1bec76b1854d5c272ad425feb7e25f2eea7760932a7fc5eec2769"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.209626 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" event={"ID":"b081f2a5-0dee-4ee9-b5af-564af14f3e69","Type":"ContainerStarted","Data":"2d073f9a00f84774929bfe12f3035d0953fa38d12c914b3b4772dafc0ba4f7b1"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.210076 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-v6c8s" podStartSLOduration=130.210051063 podStartE2EDuration="2m10.210051063s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.209290172 +0000 UTC m=+148.689593210" watchObservedRunningTime="2026-01-22 00:09:33.210051063 +0000 UTC m=+148.690354101" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.219219 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g5tbw" event={"ID":"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0","Type":"ContainerStarted","Data":"6b6bba549460b8fab15b449c21ae48c30fdfea2291258e6befbd0a1b7028458a"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.242591 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" podStartSLOduration=130.242571429 podStartE2EDuration="2m10.242571429s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.240598236 +0000 UTC m=+148.720901274" watchObservedRunningTime="2026-01-22 00:09:33.242571429 +0000 UTC m=+148.722874467" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.256269 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" event={"ID":"f017341d-0ebc-442d-8980-5e2839c89c3a","Type":"ContainerStarted","Data":"0adeda191359ed02fcee8e1fb615bf90dbe4f1e2176020edb9f41fb1bff4d71c"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.266524 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-g5tbw" podStartSLOduration=7.266508234 podStartE2EDuration="7.266508234s" podCreationTimestamp="2026-01-22 00:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.266408421 +0000 UTC m=+148.746711479" watchObservedRunningTime="2026-01-22 00:09:33.266508234 +0000 UTC m=+148.746811272" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.277914 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" event={"ID":"d516372c-17a3-438c-87f8-d2e3edfb5670","Type":"ContainerStarted","Data":"7e477237dce3d8d362b0923d1f81194c8735b1bd4101505b0bed148223a4d64f"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.278250 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.279592 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.779569255 +0000 UTC m=+149.259872293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.287521 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" event={"ID":"666d239f-66e7-48de-bc1c-6d0c1ab4f774","Type":"ContainerStarted","Data":"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.287848 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.287904 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" event={"ID":"666d239f-66e7-48de-bc1c-6d0c1ab4f774","Type":"ContainerStarted","Data":"793cca827c9b2873f7ad4aa048dad5f4ea9ba15d37da91a9ad6d65cb263ac4d3"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.290482 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" event={"ID":"e6e26d7b-9ede-44a2-97d6-f60b80b38839","Type":"ContainerStarted","Data":"d9c92c3ff0cb170c26e0facac015d9bf20f4f351ffa53b16d0c43017a709fe29"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.303412 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" event={"ID":"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8","Type":"ContainerStarted","Data":"d353e03f07a7a41bc3d099de90791faafd7b04f098a1aec35c90dbc0d440bf5b"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.304919 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" event={"ID":"025d333e-13b9-49c2-b6b8-512fd070c4db","Type":"ContainerStarted","Data":"86392a971b3e584a7ad09886118e5cd2ef3dd6f9f1c6fbfa64b322cffacfc584"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.325462 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" podStartSLOduration=131.325438151 podStartE2EDuration="2m11.325438151s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.322184383 +0000 UTC m=+148.802487421" watchObservedRunningTime="2026-01-22 00:09:33.325438151 +0000 UTC m=+148.805741189" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.358592 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ggnbw" event={"ID":"1f8f0912-b7a1-4e77-b270-4ba5e910cbf9","Type":"ContainerStarted","Data":"a2e3b8e5d4988b5cea3ee20ee1c345d01064e9a7737ae870d7c0b51d8edff65f"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.358671 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ggnbw" event={"ID":"1f8f0912-b7a1-4e77-b270-4ba5e910cbf9","Type":"ContainerStarted","Data":"f73388f84dd8842fd7fe7a70048b413da9cf7dd208884e355f09ad8530f1d345"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.360109 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.369648 4800 patch_prober.go:28] interesting pod/downloads-7954f5f757-ggnbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.369720 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ggnbw" podUID="1f8f0912-b7a1-4e77-b270-4ba5e910cbf9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.377646 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6674e8e54f18d6abccfcee50119b6016506935f6b434a6b3614413a6322ffdc9"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.382999 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" event={"ID":"092d47e3-8067-4702-bbd2-1a1bdb429b67","Type":"ContainerStarted","Data":"4acdb0fd25f4c2b16fbbb1166582443ad25df92d5d9169ebb32c8126fdf8d1ab"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.383736 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.386832 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.886814684 +0000 UTC m=+149.367117712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.395459 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ggnbw" podStartSLOduration=131.395439276 podStartE2EDuration="2m11.395439276s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.394185732 +0000 UTC m=+148.874488770" watchObservedRunningTime="2026-01-22 00:09:33.395439276 +0000 UTC m=+148.875742314" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.396316 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" event={"ID":"c86f998b-7292-4a18-a830-7d2c46dcdf2c","Type":"ContainerStarted","Data":"f97a23933edd3ade5a11d18191dc52c7fbca073126e688aa671cabbee69d4cf9"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.412104 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" event={"ID":"f5059531-28eb-4fa0-a276-3d7310a814a6","Type":"ContainerStarted","Data":"75e1941e660fda24e4d3ad61ff7d9da5f1e5cd34eeb5d57e5c797436e9e63e2c"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.412153 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" event={"ID":"f5059531-28eb-4fa0-a276-3d7310a814a6","Type":"ContainerStarted","Data":"415a3d1e84c5eadda551894428db078578262efaa3a99a5deb3934ff95e5f825"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.417666 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" event={"ID":"f695822d-7cf9-483b-bf22-ebc2bec117e8","Type":"ContainerStarted","Data":"64be63060147afaa31d6766ec21110aac63a03f051aa2de4ffc019443a1538a1"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.417700 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" event={"ID":"f695822d-7cf9-483b-bf22-ebc2bec117e8","Type":"ContainerStarted","Data":"5419ef22ea9485f30cc20eed91f070226bb05ad97d9ef87b62a926753d6a9a22"} Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.420957 4800 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r9z5h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.421017 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.462631 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-csf2g" podStartSLOduration=131.462610556 podStartE2EDuration="2m11.462610556s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.461699051 +0000 UTC m=+148.942002089" watchObservedRunningTime="2026-01-22 00:09:33.462610556 +0000 UTC m=+148.942913594" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.464958 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdggf" podStartSLOduration=131.464943619 podStartE2EDuration="2m11.464943619s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:33.435764253 +0000 UTC m=+148.916067281" watchObservedRunningTime="2026-01-22 00:09:33.464943619 +0000 UTC m=+148.945246667" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.484564 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.484716 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.98467722 +0000 UTC m=+149.464980268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.484804 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.485173 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:33.985161203 +0000 UTC m=+149.465464241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.506705 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.507673 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.517644 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.524527 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.578525 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.585501 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.585866 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.085840865 +0000 UTC m=+149.566143903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.587240 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.587335 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.587943 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.087933342 +0000 UTC m=+149.568236380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.588123 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.588608 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrn8x\" (UniqueName: \"kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.698551 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.698706 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrn8x\" (UniqueName: \"kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.698754 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.698771 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.699439 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.699522 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.199504496 +0000 UTC m=+149.679807534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.700066 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.716934 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.740037 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.740495 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.750615 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:33 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:33 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:33 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.750672 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.767400 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrn8x\" (UniqueName: \"kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x\") pod \"community-operators-2gn8v\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.771116 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.812611 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.813040 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.313027194 +0000 UTC m=+149.793330222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.899843 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.901170 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.918073 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.918218 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.418189126 +0000 UTC m=+149.898492164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.918336 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvffz\" (UniqueName: \"kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.918395 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.918421 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.918515 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:33 crc kubenswrapper[4800]: E0122 00:09:33.921010 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.420992872 +0000 UTC m=+149.901295910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:33 crc kubenswrapper[4800]: I0122 00:09:33.962969 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.020607 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021342 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x7st\" (UniqueName: \"kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021383 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021428 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvffz\" (UniqueName: \"kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021452 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021491 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.021520 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.021707 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.521690004 +0000 UTC m=+150.001993042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.022580 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.023235 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.044334 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.052089 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvffz\" (UniqueName: \"kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz\") pod \"certified-operators-hgwx4\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.109210 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.111197 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.124652 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.124736 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.124807 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x7st\" (UniqueName: \"kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.124861 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.125346 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.125666 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.625641885 +0000 UTC m=+150.105944923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.125745 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.130573 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.178228 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x7st\" (UniqueName: \"kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st\") pod \"community-operators-cb9ln\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.226682 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.226945 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.226972 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.226989 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fc7g\" (UniqueName: \"kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.227112 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.727092717 +0000 UTC m=+150.207395755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.303762 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.328438 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.328834 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.328858 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fc7g\" (UniqueName: \"kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.328965 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.329369 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.829355232 +0000 UTC m=+150.309658270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.329544 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.330430 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.346555 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.367407 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fc7g\" (UniqueName: \"kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g\") pod \"certified-operators-sxp25\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.429815 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.430111 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:34.930092695 +0000 UTC m=+150.410395723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.454367 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.463066 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" event={"ID":"29ad2446-ab96-4ca3-ba06-4a1e8d6ae6d8","Type":"ContainerStarted","Data":"2c94aa5ba793ff09eef5675045a56d5dab4af6051b44772de5e34e39c8314632"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.487449 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" event={"ID":"dbeac130-c061-4ae1-a609-8808eb849e38","Type":"ContainerStarted","Data":"c856bcf1d8c0f620ea3f79748da2a7dadaccbf29289ab16e24db1c2b4d13a12b"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.536298 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b41f1dbe16339d0f3a0c68327c11bea12dc4955d0a4548d4a7306028c4614155"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.538806 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.539780 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.039764909 +0000 UTC m=+150.520067947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.543794 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"dec22d5cd6f9a0d24116904e262e4443b2f84aa65db0140451c74ce7e029f626"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.554704 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6s75r" podStartSLOduration=131.554683471 podStartE2EDuration="2m11.554683471s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:34.503683408 +0000 UTC m=+149.983986446" watchObservedRunningTime="2026-01-22 00:09:34.554683471 +0000 UTC m=+150.034986509" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.555879 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.575707 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" event={"ID":"029287ae-ef37-46c9-b023-1d7074803fd6","Type":"ContainerStarted","Data":"73c78575b71ff22da188e87564872b156826be35e810eeafea39588b9dae181d"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.576474 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.579157 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" event={"ID":"e6de60db-1a8c-4be4-9943-c5afcd8f4566","Type":"ContainerStarted","Data":"f4fab1f540cc6dab5dddf1a2ce0c1e4345c1230528e7da153e55bbf0561b6042"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.617988 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" event={"ID":"d516372c-17a3-438c-87f8-d2e3edfb5670","Type":"ContainerStarted","Data":"b3ae518a90c71a36ce18085b65953ad3ec59d9301f84d41838f4b386dc2f8216"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.640152 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.641522 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.14150429 +0000 UTC m=+150.621807328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.669103 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" event={"ID":"909b3993-ea89-4b5a-8376-4f17044468bb","Type":"ContainerStarted","Data":"10e52f438079accd6d73a2db5804217c9b7aa9f77a4d8491b2d14e6e91e81a2c"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.706419 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" event={"ID":"025d333e-13b9-49c2-b6b8-512fd070c4db","Type":"ContainerStarted","Data":"242e8983d138a456957eca34c72d98e154d80bc71559c02b48788a2e8d7d1048"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.717560 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" podStartSLOduration=131.717535327 podStartE2EDuration="2m11.717535327s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:34.716337036 +0000 UTC m=+150.196640074" watchObservedRunningTime="2026-01-22 00:09:34.717535327 +0000 UTC m=+150.197838365" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.732850 4800 generic.go:334] "Generic (PLEG): container finished" podID="c9693007-9dd0-46b0-9f2d-b75421bfcde9" containerID="dbf2fef4f8c58d7f543e7d7fb520b02f4b0b8c3545899f8c4c90964e4b6f4123" exitCode=0 Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.732982 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" event={"ID":"c9693007-9dd0-46b0-9f2d-b75421bfcde9","Type":"ContainerDied","Data":"dbf2fef4f8c58d7f543e7d7fb520b02f4b0b8c3545899f8c4c90964e4b6f4123"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.744171 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.744495 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.244485204 +0000 UTC m=+150.724788242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.749231 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:34 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:34 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:34 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.749301 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.771983 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" event={"ID":"092d47e3-8067-4702-bbd2-1a1bdb429b67","Type":"ContainerStarted","Data":"3cda8b2386d9d805742512c9448896e9e982d972aa7d233e54a0ec3c86895d88"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.773375 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.775765 4800 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fzt4x container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.775814 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" podUID="092d47e3-8067-4702-bbd2-1a1bdb429b67" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.855455 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.858847 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.358814263 +0000 UTC m=+150.839117301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.872979 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-msgt9" podStartSLOduration=131.872952913 podStartE2EDuration="2m11.872952913s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:34.811037076 +0000 UTC m=+150.291340114" watchObservedRunningTime="2026-01-22 00:09:34.872952913 +0000 UTC m=+150.353255951" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.928474 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" podStartSLOduration=131.928453449 podStartE2EDuration="2m11.928453449s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:34.927347389 +0000 UTC m=+150.407650427" watchObservedRunningTime="2026-01-22 00:09:34.928453449 +0000 UTC m=+150.408756487" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.970850 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.971619 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" event={"ID":"f017341d-0ebc-442d-8980-5e2839c89c3a","Type":"ContainerStarted","Data":"367173ab711b13c8b6d52a0d82f28da45004f1f926e9535febe3836e979b628d"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.971704 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" event={"ID":"b081f2a5-0dee-4ee9-b5af-564af14f3e69","Type":"ContainerStarted","Data":"7370814377f0904f5e151e3155db9994544d1524e995ac48d582cc0d139a79c6"} Jan 22 00:09:34 crc kubenswrapper[4800]: E0122 00:09:34.973399 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.473385289 +0000 UTC m=+150.953688327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.993525 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" event={"ID":"c86f998b-7292-4a18-a830-7d2c46dcdf2c","Type":"ContainerStarted","Data":"bd6721db9b3a0450c8354a6f0f338f33e83465314a7424c025da3d967c0b162d"} Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.995361 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:34 crc kubenswrapper[4800]: I0122 00:09:34.999567 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-shzt2" podStartSLOduration=132.999539874 podStartE2EDuration="2m12.999539874s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:34.99830693 +0000 UTC m=+150.478609968" watchObservedRunningTime="2026-01-22 00:09:34.999539874 +0000 UTC m=+150.479842912" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.000065 4800 patch_prober.go:28] interesting pod/console-operator-58897d9998-6t2gz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.000139 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" podUID="c86f998b-7292-4a18-a830-7d2c46dcdf2c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.071789 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.072167 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.57214971 +0000 UTC m=+151.052452748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.113597 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" event={"ID":"eee929d7-38fc-464c-9ca1-a32f6e1771b5","Type":"ContainerStarted","Data":"55e6d1d850818f9ba1e711a747ec4f7dde5e667c55ca56df2b84855737dc4ba7"} Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.161281 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" event={"ID":"ce65736c-5fd3-4d03-9cf2-e8759f560551","Type":"ContainerStarted","Data":"ac1922b553e117ca47f6e864e985253e9cb55f5343591822499a443a2e9c914a"} Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.173412 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.173794 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.673779066 +0000 UTC m=+151.154082104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.201946 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"acfaf341126fcba418b17902797235df55b2e9c53fe8e1318b004b2e9aa1b3b5"} Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.230484 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" podStartSLOduration=133.230462974 podStartE2EDuration="2m13.230462974s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:35.228993834 +0000 UTC m=+150.709296872" watchObservedRunningTime="2026-01-22 00:09:35.230462974 +0000 UTC m=+150.710766002" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.255043 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" event={"ID":"e6e26d7b-9ede-44a2-97d6-f60b80b38839","Type":"ContainerStarted","Data":"5a244b3e9711e06cf1d31c8b907ff9031ea2a5b10408d26308fe230bdb8ab457"} Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.264048 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g5tbw" event={"ID":"7f1c1bd7-3df1-4da2-bb0c-e30a5aeb7fc0","Type":"ContainerStarted","Data":"d405a858ff23488b63d426b55b0cc1b487ffe74b50e0acaff4aca61bc9347e37"} Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.274294 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.276774 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.77675001 +0000 UTC m=+151.257053118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.334505 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nzbgw" podStartSLOduration=133.334481995 podStartE2EDuration="2m13.334481995s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:35.334024053 +0000 UTC m=+150.814327091" watchObservedRunningTime="2026-01-22 00:09:35.334481995 +0000 UTC m=+150.814785033" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.335345 4800 patch_prober.go:28] interesting pod/downloads-7954f5f757-ggnbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.335396 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ggnbw" podUID="1f8f0912-b7a1-4e77-b270-4ba5e910cbf9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.359422 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n9cq8" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.386934 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.387690 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.887676008 +0000 UTC m=+151.367979046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.467042 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" podStartSLOduration=133.467019635 podStartE2EDuration="2m13.467019635s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:35.466006238 +0000 UTC m=+150.946309276" watchObservedRunningTime="2026-01-22 00:09:35.467019635 +0000 UTC m=+150.947322673" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.490006 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.496170 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:35.996125609 +0000 UTC m=+151.476428647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.585664 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.586987 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.594048 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.594759 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.094746966 +0000 UTC m=+151.575050004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.610281 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.671131 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.696936 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.709622 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.209578449 +0000 UTC m=+151.689881477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.710345 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.710462 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kbcg\" (UniqueName: \"kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.710607 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.710986 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.711725 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.211694565 +0000 UTC m=+151.691997603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.730999 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:35 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:35 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:35 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.731081 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.811761 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.816290 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.816399 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.316369525 +0000 UTC m=+151.796672563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.821665 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.821780 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kbcg\" (UniqueName: \"kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.821915 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.822002 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.822469 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.322454109 +0000 UTC m=+151.802757137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.822923 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.823394 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.826795 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.847210 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zsr9k" podStartSLOduration=132.847193335 podStartE2EDuration="2m12.847193335s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:35.787422595 +0000 UTC m=+151.267725633" watchObservedRunningTime="2026-01-22 00:09:35.847193335 +0000 UTC m=+151.327496373" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.897160 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kbcg\" (UniqueName: \"kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg\") pod \"redhat-marketplace-jpdkg\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.926045 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:35 crc kubenswrapper[4800]: E0122 00:09:35.926377 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.426356388 +0000 UTC m=+151.906659426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.934564 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.935937 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.975960 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:09:35 crc kubenswrapper[4800]: I0122 00:09:35.983141 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sjdlp" podStartSLOduration=132.983121556 podStartE2EDuration="2m12.983121556s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:35.951454514 +0000 UTC m=+151.431757542" watchObservedRunningTime="2026-01-22 00:09:35.983121556 +0000 UTC m=+151.463424594" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.032134 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.032193 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.032267 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgcj5\" (UniqueName: \"kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.032309 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.032791 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.532774874 +0000 UTC m=+152.013077912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.038081 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.052216 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.136531 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.137164 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.137400 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgcj5\" (UniqueName: \"kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.137528 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.138057 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.138198 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.638178563 +0000 UTC m=+152.118481601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.138529 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.193463 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgcj5\" (UniqueName: \"kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5\") pod \"redhat-marketplace-xdbc4\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.239779 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.240107 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.740090458 +0000 UTC m=+152.220393496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.342338 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.343170 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.843140053 +0000 UTC m=+152.323443091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.343359 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.343956 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.843948646 +0000 UTC m=+152.324251674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.369497 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.395655 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" event={"ID":"025d333e-13b9-49c2-b6b8-512fd070c4db","Type":"ContainerStarted","Data":"50b1076fbb5d1eaabf84b63f538885cef6b73ea6c041d4ccf36c932411787f89"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.445293 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.445662 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:36.945640935 +0000 UTC m=+152.425943973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.455740 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" event={"ID":"c9693007-9dd0-46b0-9f2d-b75421bfcde9","Type":"ContainerStarted","Data":"57c337b6eb604fcaa885d7f874fe988ea14842bdc857bac1d9af0d4c8a3358c8"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.456983 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.483754 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" podStartSLOduration=134.483728241 podStartE2EDuration="2m14.483728241s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:36.479841246 +0000 UTC m=+151.960144284" watchObservedRunningTime="2026-01-22 00:09:36.483728241 +0000 UTC m=+151.964031279" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.557866 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.559454 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.05943998 +0000 UTC m=+152.539743018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.572386 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" event={"ID":"85468a31-5db8-44d1-8027-afab2b6f8e02","Type":"ContainerStarted","Data":"ecee47ee8be8f5c5a516bc6a98dda4354ad48a939f15d3b4cc783d0e0f68a9d6"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.586664 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" event={"ID":"d516372c-17a3-438c-87f8-d2e3edfb5670","Type":"ContainerStarted","Data":"9a535a7b6294a82c93cd989db1671cce82f1434556c93ed0c04523e2b048dadb"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.599112 4800 generic.go:334] "Generic (PLEG): container finished" podID="6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7" containerID="cab3950f520fc35b4fe0499a2a0d9dfba1c42a16bd1d926c808d85d5275e1ea7" exitCode=0 Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.599358 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" event={"ID":"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7","Type":"ContainerDied","Data":"cab3950f520fc35b4fe0499a2a0d9dfba1c42a16bd1d926c808d85d5275e1ea7"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.601366 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" podStartSLOduration=133.601344589 podStartE2EDuration="2m13.601344589s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:36.599002435 +0000 UTC m=+152.079305483" watchObservedRunningTime="2026-01-22 00:09:36.601344589 +0000 UTC m=+152.081647637" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.612396 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ff9facc-6840-481a-80d2-c698c359564c" containerID="8bc90665985f86ee2cba1121e3c10263afb6c10e477e9c227b206a0c659160ad" exitCode=0 Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.612552 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerDied","Data":"8bc90665985f86ee2cba1121e3c10263afb6c10e477e9c227b206a0c659160ad"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.612600 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerStarted","Data":"6b2bd2d3c8488e6534ed07035f49dc23546ab0da847a6a07739985da0d1cf14c"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.625263 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.657104 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-shdlf" podStartSLOduration=133.65708005 podStartE2EDuration="2m13.65708005s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:36.639999779 +0000 UTC m=+152.120302817" watchObservedRunningTime="2026-01-22 00:09:36.65708005 +0000 UTC m=+152.137383088" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.660759 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.664926 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.164869939 +0000 UTC m=+152.645172977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.679636 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6l6mb" event={"ID":"25c87458-5680-441e-8d4f-8f8b3d0ea5d5","Type":"ContainerStarted","Data":"b87afe71a934fdc46b372e493f0c423282232cb2a20b531d6d4c0726f7424f45"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.679697 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6l6mb" event={"ID":"25c87458-5680-441e-8d4f-8f8b3d0ea5d5","Type":"ContainerStarted","Data":"cefdd268cb9f1aefd2eb82e16a7e1a4c48064c652aba67e766c8dd3b0679e818"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.680068 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.732149 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.742674 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.749839 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" event={"ID":"dbeac130-c061-4ae1-a609-8808eb849e38","Type":"ContainerStarted","Data":"e6e383824ce687a52a55772ef1ab86c82cf84c7f74e7ec35da403a6b01424622"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.751354 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.756543 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:36 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:36 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:36 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.757018 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.776049 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerStarted","Data":"e4d666f683dc23e9f8a4bdeb3e12804863b3114690e109ec9a08a40db28997d5"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.800473 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.800528 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk42q\" (UniqueName: \"kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.800559 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.800593 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.801646 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.301625323 +0000 UTC m=+152.781928361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.850257 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6l6mb" podStartSLOduration=10.850231123 podStartE2EDuration="10.850231123s" podCreationTimestamp="2026-01-22 00:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:36.806448983 +0000 UTC m=+152.286752011" watchObservedRunningTime="2026-01-22 00:09:36.850231123 +0000 UTC m=+152.330534161" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.854580 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.860312 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lvwxh" event={"ID":"b081f2a5-0dee-4ee9-b5af-564af14f3e69","Type":"ContainerStarted","Data":"289e97765746639b819adee16143c30749afc36118e0a28314c1d160f317729b"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.886563 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.902603 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.902844 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.902901 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk42q\" (UniqueName: \"kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.902938 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.903460 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: E0122 00:09:36.903538 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.403523588 +0000 UTC m=+152.883826626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.905563 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.947478 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cv9lb" podStartSLOduration=134.947453221 podStartE2EDuration="2m14.947453221s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:36.947358479 +0000 UTC m=+152.427661517" watchObservedRunningTime="2026-01-22 00:09:36.947453221 +0000 UTC m=+152.427756259" Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.954669 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerStarted","Data":"b430618eda7d5a2619933a352aef6706b7ac948a1e5473c18d4f988b1b4b659b"} Jan 22 00:09:36 crc kubenswrapper[4800]: I0122 00:09:36.993356 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerStarted","Data":"ea42d617f2dd2f5139f31b3c8057f8e41d573e19ef8eceb3bd915b7c77f73424"} Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.010381 4800 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.012564 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.012866 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.512853003 +0000 UTC m=+152.993156041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.037756 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk42q\" (UniqueName: \"kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q\") pod \"redhat-operators-n28k4\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.054397 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" event={"ID":"ce65736c-5fd3-4d03-9cf2-e8759f560551","Type":"ContainerStarted","Data":"3446104f5f320da198c127be20eb0793b5488288ae2e62a41bedc1d255be2259"} Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.058251 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.085195 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6t2gz" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.091769 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fzt4x" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.111973 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.113680 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.115993 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.61596592 +0000 UTC m=+153.096268958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.126057 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.132686 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.146598 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.149714 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.175480 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-v5n9s" podStartSLOduration=134.175463133 podStartE2EDuration="2m14.175463133s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:37.175262967 +0000 UTC m=+152.655566005" watchObservedRunningTime="2026-01-22 00:09:37.175463133 +0000 UTC m=+152.655766171" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.218257 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.218583 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.218678 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw44b\" (UniqueName: \"kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.218771 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.219187 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.71917203 +0000 UTC m=+153.199475068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.320809 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.321095 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.821072135 +0000 UTC m=+153.301375173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.321500 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.321587 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.321702 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw44b\" (UniqueName: \"kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.321824 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.322613 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.322833 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.822811791 +0000 UTC m=+153.303114829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.323212 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.356535 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw44b\" (UniqueName: \"kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b\") pod \"redhat-operators-577ft\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.424225 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.424359 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.924327896 +0000 UTC m=+153.404630934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.425151 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.425608 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:37.925592509 +0000 UTC m=+153.405895547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.480366 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.527550 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.527918 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.027901306 +0000 UTC m=+153.508204344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.628841 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.629219 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.129199164 +0000 UTC m=+153.609502202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.676454 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:09:37 crc kubenswrapper[4800]: W0122 00:09:37.706260 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6935c82f_f42c_44c6_90c5_968f24feed80.slice/crio-13ea5a7aceb71b131dd37e2acc87eb726edc23fc62d3073be7207d8bd22efaef WatchSource:0}: Error finding container 13ea5a7aceb71b131dd37e2acc87eb726edc23fc62d3073be7207d8bd22efaef: Status 404 returned error can't find the container with id 13ea5a7aceb71b131dd37e2acc87eb726edc23fc62d3073be7207d8bd22efaef Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.729935 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.730363 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.230343448 +0000 UTC m=+153.710646496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.730547 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:37 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:37 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:37 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.730581 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.831801 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.832369 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.332354836 +0000 UTC m=+153.812657874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.934863 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.935711 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.435678769 +0000 UTC m=+153.915981797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.937442 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:37 crc kubenswrapper[4800]: E0122 00:09:37.937933 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 00:09:38.437914659 +0000 UTC m=+153.918217697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2qvn4" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.984791 4800 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-22T00:09:37.010408487Z","Handler":null,"Name":""} Jan 22 00:09:37 crc kubenswrapper[4800]: I0122 00:09:37.989181 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.004551 4800 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.004591 4800 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.038515 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.043074 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.081483 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerStarted","Data":"bfe3fdbf911965fc3397b65841b93fce985432a0caea838089f3cb0f6d3c7538"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.086762 4800 generic.go:334] "Generic (PLEG): container finished" podID="0f5bfc30-ad0a-486a-baad-9dc8249f7170" containerID="75627e11053cb23ff4dacefd3832ce37a3419f699288840153f5a35807fe7040" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.086878 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" event={"ID":"0f5bfc30-ad0a-486a-baad-9dc8249f7170","Type":"ContainerDied","Data":"75627e11053cb23ff4dacefd3832ce37a3419f699288840153f5a35807fe7040"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.089417 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerStarted","Data":"f2ac286b7cca96fa19b30e06f29b2e412f39b4143f4a2217cefa65950d4d56ea"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.089467 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerStarted","Data":"13ea5a7aceb71b131dd37e2acc87eb726edc23fc62d3073be7207d8bd22efaef"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.091787 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerID="f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.091910 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerDied","Data":"f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.100711 4800 generic.go:334] "Generic (PLEG): container finished" podID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerID="67a397d4096cc85de2f875409edf4b3335c224853dd6455f126f3fc9af7f9135" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.100862 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerDied","Data":"67a397d4096cc85de2f875409edf4b3335c224853dd6455f126f3fc9af7f9135"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.100931 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerStarted","Data":"7a24fdd8de936c9f24516bcdee24b159ae341d265ad234eb7fe6789107edf27b"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.114267 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" event={"ID":"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7","Type":"ContainerStarted","Data":"78d39c0de3ca1002cca37665a53193a38928bc02fe54ede801b88e64a3a66337"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.114316 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" event={"ID":"6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7","Type":"ContainerStarted","Data":"ce87c3a1989958a89c9a598431d505b0f2aeeea2f48371782ce28a1b5d2b03f2"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.114462 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x99cx" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.117048 4800 generic.go:334] "Generic (PLEG): container finished" podID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerID="8742b7bcf620717d57cd5b314c0cc37cd7fd28c3bc9a4f1d6b1a8cfef674fc39" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.117118 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerDied","Data":"8742b7bcf620717d57cd5b314c0cc37cd7fd28c3bc9a4f1d6b1a8cfef674fc39"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.134402 4800 generic.go:334] "Generic (PLEG): container finished" podID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerID="1b6b25536c8c149379b7c82a1a66bc968a2f4d2fe2bb64cce5e0d8634afc426b" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.134601 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerDied","Data":"1b6b25536c8c149379b7c82a1a66bc968a2f4d2fe2bb64cce5e0d8634afc426b"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.138647 4800 generic.go:334] "Generic (PLEG): container finished" podID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerID="081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745" exitCode=0 Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.140859 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.141195 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerDied","Data":"081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.142999 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerStarted","Data":"4cf6dc30ccdcfc5f1b177941a7b4d15e4c2015e530871bdaffde926d566d8367"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.147984 4800 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.148057 4800 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.159657 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" event={"ID":"025d333e-13b9-49c2-b6b8-512fd070c4db","Type":"ContainerStarted","Data":"4af1f94b12f2bf0ac19d35098560e1c1d98faebddc4c0d734686e174784c919d"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.159703 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" event={"ID":"025d333e-13b9-49c2-b6b8-512fd070c4db","Type":"ContainerStarted","Data":"96da48098c1ea4a02d1fc890d0f445cc7719dec526f55b9970bffd04ccb523ba"} Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.222003 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2qvn4\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.252124 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" podStartSLOduration=136.252084842 podStartE2EDuration="2m16.252084842s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:38.240270613 +0000 UTC m=+153.720573661" watchObservedRunningTime="2026-01-22 00:09:38.252084842 +0000 UTC m=+153.732387880" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.291309 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7gmzt" podStartSLOduration=12.291283608 podStartE2EDuration="12.291283608s" podCreationTimestamp="2026-01-22 00:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:38.290442575 +0000 UTC m=+153.770745613" watchObservedRunningTime="2026-01-22 00:09:38.291283608 +0000 UTC m=+153.771586646" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.317705 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.626426 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.731249 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:38 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:38 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:38 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.731396 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.852479 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 22 00:09:38 crc kubenswrapper[4800]: I0122 00:09:38.902031 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.029730 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.029853 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.050064 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.050466 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.054504 4800 patch_prober.go:28] interesting pod/console-f9d7485db-q6r6s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.054571 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q6r6s" podUID="5543f84b-8bd9-4464-9ee7-68ef8ec50985" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.069596 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.178640 4800 generic.go:334] "Generic (PLEG): container finished" podID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerID="dc1c5fb08514854fb58063a4a3615cf17b597aeb7ae8255de3db7235943d47ac" exitCode=0 Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.178873 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerDied","Data":"dc1c5fb08514854fb58063a4a3615cf17b597aeb7ae8255de3db7235943d47ac"} Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.195763 4800 generic.go:334] "Generic (PLEG): container finished" podID="6935c82f-f42c-44c6-90c5-968f24feed80" containerID="f2ac286b7cca96fa19b30e06f29b2e412f39b4143f4a2217cefa65950d4d56ea" exitCode=0 Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.195868 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerDied","Data":"f2ac286b7cca96fa19b30e06f29b2e412f39b4143f4a2217cefa65950d4d56ea"} Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.213213 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" event={"ID":"4b3e309b-172b-4da2-91f5-8be5152cae2e","Type":"ContainerStarted","Data":"eec5d82aa11ecb1c9de87217a9fd022dccbcb6292efea48048afcfd85d7a6a62"} Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.227470 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fmb8f" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.259954 4800 patch_prober.go:28] interesting pod/downloads-7954f5f757-ggnbw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.260011 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ggnbw" podUID="1f8f0912-b7a1-4e77-b270-4ba5e910cbf9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.260157 4800 patch_prober.go:28] interesting pod/downloads-7954f5f757-ggnbw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.260294 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ggnbw" podUID="1f8f0912-b7a1-4e77-b270-4ba5e910cbf9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.355787 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.356156 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.380587 4800 patch_prober.go:28] interesting pod/apiserver-76f77b778f-zn8zh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]log ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]etcd ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/generic-apiserver-start-informers ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/max-in-flight-filter ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 22 00:09:39 crc kubenswrapper[4800]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 22 00:09:39 crc kubenswrapper[4800]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/project.openshift.io-projectcache ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-startinformers ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 22 00:09:39 crc kubenswrapper[4800]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 22 00:09:39 crc kubenswrapper[4800]: livez check failed Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.380653 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" podUID="6d9bb78d-8a6f-45a8-a6b1-2382eafd27a7" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.687120 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.688554 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.717627 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.717961 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.724209 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.724831 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.732450 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:39 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:39 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:39 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.732509 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.744910 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.819552 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrxxn\" (UniqueName: \"kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn\") pod \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.819606 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume\") pod \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.819774 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume\") pod \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\" (UID: \"0f5bfc30-ad0a-486a-baad-9dc8249f7170\") " Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.820058 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.820120 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.822527 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume" (OuterVolumeSpecName: "config-volume") pod "0f5bfc30-ad0a-486a-baad-9dc8249f7170" (UID: "0f5bfc30-ad0a-486a-baad-9dc8249f7170"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.835432 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn" (OuterVolumeSpecName: "kube-api-access-jrxxn") pod "0f5bfc30-ad0a-486a-baad-9dc8249f7170" (UID: "0f5bfc30-ad0a-486a-baad-9dc8249f7170"). InnerVolumeSpecName "kube-api-access-jrxxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.859007 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0f5bfc30-ad0a-486a-baad-9dc8249f7170" (UID: "0f5bfc30-ad0a-486a-baad-9dc8249f7170"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.921599 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.921712 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.921831 4800 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f5bfc30-ad0a-486a-baad-9dc8249f7170-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.921844 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrxxn\" (UniqueName: \"kubernetes.io/projected/0f5bfc30-ad0a-486a-baad-9dc8249f7170-kube-api-access-jrxxn\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.921869 4800 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f5bfc30-ad0a-486a-baad-9dc8249f7170-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.922256 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:39 crc kubenswrapper[4800]: I0122 00:09:39.974939 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.066271 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.240983 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" event={"ID":"0f5bfc30-ad0a-486a-baad-9dc8249f7170","Type":"ContainerDied","Data":"18496040af9a2f886482d376fb89e2d1dfeb519b8c57a6757a03fefbbc44540a"} Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.241048 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18496040af9a2f886482d376fb89e2d1dfeb519b8c57a6757a03fefbbc44540a" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.241172 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.262078 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" event={"ID":"4b3e309b-172b-4da2-91f5-8be5152cae2e","Type":"ContainerStarted","Data":"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177"} Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.262699 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.328861 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" podStartSLOduration=137.328824919 podStartE2EDuration="2m17.328824919s" podCreationTimestamp="2026-01-22 00:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:40.326731604 +0000 UTC m=+155.807034642" watchObservedRunningTime="2026-01-22 00:09:40.328824919 +0000 UTC m=+155.809127977" Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.629276 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.728472 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:40 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:40 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:40 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:40 crc kubenswrapper[4800]: I0122 00:09:40.728560 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.296251 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"978eb551-f0e5-4448-a66a-5b8256b93312","Type":"ContainerStarted","Data":"c7284d9f6c8bf48ec410625fe14f02007a86b40100f2c195f6516662b4052a2d"} Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.727831 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:41 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:41 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:41 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.728501 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.837041 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 00:09:41 crc kubenswrapper[4800]: E0122 00:09:41.837606 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f5bfc30-ad0a-486a-baad-9dc8249f7170" containerName="collect-profiles" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.837658 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f5bfc30-ad0a-486a-baad-9dc8249f7170" containerName="collect-profiles" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.837864 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f5bfc30-ad0a-486a-baad-9dc8249f7170" containerName="collect-profiles" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.839103 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.843491 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.843494 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.859302 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.859476 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.879251 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.960951 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.961072 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.961543 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:41 crc kubenswrapper[4800]: I0122 00:09:41.987422 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.169097 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.322025 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"978eb551-f0e5-4448-a66a-5b8256b93312","Type":"ContainerStarted","Data":"a9a48f8dda4f34694e641f09039e75307951df230dacb10565da3880fa02f96b"} Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.345653 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.345628582 podStartE2EDuration="3.345628582s" podCreationTimestamp="2026-01-22 00:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:42.339879498 +0000 UTC m=+157.820182536" watchObservedRunningTime="2026-01-22 00:09:42.345628582 +0000 UTC m=+157.825931630" Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.484488 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 00:09:42 crc kubenswrapper[4800]: W0122 00:09:42.516821 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9db46e68_7be5_42fd_b7a4_9ee12dc0d113.slice/crio-4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf WatchSource:0}: Error finding container 4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf: Status 404 returned error can't find the container with id 4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.729200 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:42 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:42 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:42 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:42 crc kubenswrapper[4800]: I0122 00:09:42.729637 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:43 crc kubenswrapper[4800]: I0122 00:09:43.342603 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9db46e68-7be5-42fd-b7a4-9ee12dc0d113","Type":"ContainerStarted","Data":"4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf"} Jan 22 00:09:43 crc kubenswrapper[4800]: I0122 00:09:43.347727 4800 generic.go:334] "Generic (PLEG): container finished" podID="978eb551-f0e5-4448-a66a-5b8256b93312" containerID="a9a48f8dda4f34694e641f09039e75307951df230dacb10565da3880fa02f96b" exitCode=0 Jan 22 00:09:43 crc kubenswrapper[4800]: I0122 00:09:43.347780 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"978eb551-f0e5-4448-a66a-5b8256b93312","Type":"ContainerDied","Data":"a9a48f8dda4f34694e641f09039e75307951df230dacb10565da3880fa02f96b"} Jan 22 00:09:43 crc kubenswrapper[4800]: I0122 00:09:43.730488 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:43 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:43 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:43 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:43 crc kubenswrapper[4800]: I0122 00:09:43.730571 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.363004 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.368400 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-zn8zh" Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.396366 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9db46e68-7be5-42fd-b7a4-9ee12dc0d113","Type":"ContainerStarted","Data":"8a5b80ed02631784578fb8040c7c32360e3d5921be1a05967589a5f240c1787a"} Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.562569 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6l6mb" Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.596956 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.596934492 podStartE2EDuration="3.596934492s" podCreationTimestamp="2026-01-22 00:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:09:44.500586348 +0000 UTC m=+159.980889386" watchObservedRunningTime="2026-01-22 00:09:44.596934492 +0000 UTC m=+160.077237530" Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.727738 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:44 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:44 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:44 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:44 crc kubenswrapper[4800]: I0122 00:09:44.727804 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.424213 4800 generic.go:334] "Generic (PLEG): container finished" podID="9db46e68-7be5-42fd-b7a4-9ee12dc0d113" containerID="8a5b80ed02631784578fb8040c7c32360e3d5921be1a05967589a5f240c1787a" exitCode=0 Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.424854 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9db46e68-7be5-42fd-b7a4-9ee12dc0d113","Type":"ContainerDied","Data":"8a5b80ed02631784578fb8040c7c32360e3d5921be1a05967589a5f240c1787a"} Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.727837 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:45 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:45 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:45 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.727941 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.738444 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:45 crc kubenswrapper[4800]: I0122 00:09:45.768039 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca19c294-fff4-4fe6-a1a4-25e04d0624b3-metrics-certs\") pod \"network-metrics-daemon-svkb9\" (UID: \"ca19c294-fff4-4fe6-a1a4-25e04d0624b3\") " pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:46 crc kubenswrapper[4800]: I0122 00:09:46.063159 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svkb9" Jan 22 00:09:46 crc kubenswrapper[4800]: I0122 00:09:46.728396 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:46 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:46 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:46 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:46 crc kubenswrapper[4800]: I0122 00:09:46.729069 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:47 crc kubenswrapper[4800]: I0122 00:09:47.727462 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:47 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:47 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:47 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:47 crc kubenswrapper[4800]: I0122 00:09:47.727577 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:48 crc kubenswrapper[4800]: I0122 00:09:48.728246 4800 patch_prober.go:28] interesting pod/router-default-5444994796-ffvfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 00:09:48 crc kubenswrapper[4800]: [-]has-synced failed: reason withheld Jan 22 00:09:48 crc kubenswrapper[4800]: [+]process-running ok Jan 22 00:09:48 crc kubenswrapper[4800]: healthz check failed Jan 22 00:09:48 crc kubenswrapper[4800]: I0122 00:09:48.728497 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ffvfb" podUID="a5e37c96-49b2-4450-898d-d01847863fff" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.030482 4800 patch_prober.go:28] interesting pod/console-f9d7485db-q6r6s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.030543 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q6r6s" podUID="5543f84b-8bd9-4464-9ee7-68ef8ec50985" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.276007 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ggnbw" Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.451548 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.451629 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.728286 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:49 crc kubenswrapper[4800]: I0122 00:09:49.733960 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ffvfb" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.272396 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.384906 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access\") pod \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.384992 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir\") pod \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\" (UID: \"9db46e68-7be5-42fd-b7a4-9ee12dc0d113\") " Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.385243 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9db46e68-7be5-42fd-b7a4-9ee12dc0d113" (UID: "9db46e68-7be5-42fd-b7a4-9ee12dc0d113"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.406383 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9db46e68-7be5-42fd-b7a4-9ee12dc0d113" (UID: "9db46e68-7be5-42fd-b7a4-9ee12dc0d113"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.486278 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.486319 4800 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9db46e68-7be5-42fd-b7a4-9ee12dc0d113-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.533077 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9db46e68-7be5-42fd-b7a4-9ee12dc0d113","Type":"ContainerDied","Data":"4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf"} Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.533123 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d0f4917abceefab445405ed194d8386290415e5c753db187790f54bbbd865bf" Jan 22 00:09:52 crc kubenswrapper[4800]: I0122 00:09:52.533151 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.728238 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.802195 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access\") pod \"978eb551-f0e5-4448-a66a-5b8256b93312\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.802705 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir\") pod \"978eb551-f0e5-4448-a66a-5b8256b93312\" (UID: \"978eb551-f0e5-4448-a66a-5b8256b93312\") " Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.802951 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "978eb551-f0e5-4448-a66a-5b8256b93312" (UID: "978eb551-f0e5-4448-a66a-5b8256b93312"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.803746 4800 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/978eb551-f0e5-4448-a66a-5b8256b93312-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.810117 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "978eb551-f0e5-4448-a66a-5b8256b93312" (UID: "978eb551-f0e5-4448-a66a-5b8256b93312"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:09:53 crc kubenswrapper[4800]: I0122 00:09:53.904919 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/978eb551-f0e5-4448-a66a-5b8256b93312-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:09:54 crc kubenswrapper[4800]: I0122 00:09:54.122736 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-svkb9"] Jan 22 00:09:54 crc kubenswrapper[4800]: W0122 00:09:54.132126 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca19c294_fff4_4fe6_a1a4_25e04d0624b3.slice/crio-422ca985cdf211d052324364d19c4d13810697f02161cf764ea35ddf4c7cf260 WatchSource:0}: Error finding container 422ca985cdf211d052324364d19c4d13810697f02161cf764ea35ddf4c7cf260: Status 404 returned error can't find the container with id 422ca985cdf211d052324364d19c4d13810697f02161cf764ea35ddf4c7cf260 Jan 22 00:09:54 crc kubenswrapper[4800]: I0122 00:09:54.546762 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svkb9" event={"ID":"ca19c294-fff4-4fe6-a1a4-25e04d0624b3","Type":"ContainerStarted","Data":"422ca985cdf211d052324364d19c4d13810697f02161cf764ea35ddf4c7cf260"} Jan 22 00:09:54 crc kubenswrapper[4800]: I0122 00:09:54.548553 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"978eb551-f0e5-4448-a66a-5b8256b93312","Type":"ContainerDied","Data":"c7284d9f6c8bf48ec410625fe14f02007a86b40100f2c195f6516662b4052a2d"} Jan 22 00:09:54 crc kubenswrapper[4800]: I0122 00:09:54.548597 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7284d9f6c8bf48ec410625fe14f02007a86b40100f2c195f6516662b4052a2d" Jan 22 00:09:54 crc kubenswrapper[4800]: I0122 00:09:54.548622 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 00:09:55 crc kubenswrapper[4800]: I0122 00:09:55.555205 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svkb9" event={"ID":"ca19c294-fff4-4fe6-a1a4-25e04d0624b3","Type":"ContainerStarted","Data":"9749ef8f22da4d4242bda196c74ff15e56227470bab249cd819fbbc563593c56"} Jan 22 00:09:58 crc kubenswrapper[4800]: I0122 00:09:58.327218 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:09:59 crc kubenswrapper[4800]: I0122 00:09:59.035730 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:09:59 crc kubenswrapper[4800]: I0122 00:09:59.040526 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-q6r6s" Jan 22 00:10:01 crc kubenswrapper[4800]: I0122 00:10:01.589174 4800 generic.go:334] "Generic (PLEG): container finished" podID="d6adcf93-e68c-4766-85f9-948c65a0d397" containerID="7ab6ee4a6dbb795fbc0e6fb0d94d0ee039251f7180fa495ab2a5cb11c82d6b19" exitCode=0 Jan 22 00:10:01 crc kubenswrapper[4800]: I0122 00:10:01.589282 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29484000-kr6nl" event={"ID":"d6adcf93-e68c-4766-85f9-948c65a0d397","Type":"ContainerDied","Data":"7ab6ee4a6dbb795fbc0e6fb0d94d0ee039251f7180fa495ab2a5cb11c82d6b19"} Jan 22 00:10:06 crc kubenswrapper[4800]: E0122 00:10:06.414384 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 00:10:06 crc kubenswrapper[4800]: E0122 00:10:06.415450 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xk42q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-n28k4_openshift-marketplace(6935c82f-f42c-44c6-90c5-968f24feed80): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:06 crc kubenswrapper[4800]: E0122 00:10:06.416709 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-n28k4" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" Jan 22 00:10:07 crc kubenswrapper[4800]: E0122 00:10:07.007189 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 00:10:07 crc kubenswrapper[4800]: E0122 00:10:07.007448 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gw44b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-577ft_openshift-marketplace(c8ee1ff3-44c1-4993-91e7-bb5ba795ea04): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:07 crc kubenswrapper[4800]: E0122 00:10:07.009542 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-577ft" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" Jan 22 00:10:09 crc kubenswrapper[4800]: I0122 00:10:09.242241 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rlm9b" Jan 22 00:10:11 crc kubenswrapper[4800]: I0122 00:10:11.981185 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 00:10:12 crc kubenswrapper[4800]: E0122 00:10:12.105845 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-577ft" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" Jan 22 00:10:12 crc kubenswrapper[4800]: E0122 00:10:12.105871 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-n28k4" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" Jan 22 00:10:17 crc kubenswrapper[4800]: E0122 00:10:17.205608 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 00:10:17 crc kubenswrapper[4800]: E0122 00:10:17.206131 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvffz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hgwx4_openshift-marketplace(19d5c725-0286-404e-9f66-a22ed5c2649c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:17 crc kubenswrapper[4800]: E0122 00:10:17.207276 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hgwx4" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" Jan 22 00:10:18 crc kubenswrapper[4800]: E0122 00:10:18.059248 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 00:10:18 crc kubenswrapper[4800]: E0122 00:10:18.059504 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9fc7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-sxp25_openshift-marketplace(0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:18 crc kubenswrapper[4800]: E0122 00:10:18.061053 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-sxp25" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.045389 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 00:10:19 crc kubenswrapper[4800]: E0122 00:10:19.045812 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978eb551-f0e5-4448-a66a-5b8256b93312" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.045825 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="978eb551-f0e5-4448-a66a-5b8256b93312" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: E0122 00:10:19.045870 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db46e68-7be5-42fd-b7a4-9ee12dc0d113" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.045877 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db46e68-7be5-42fd-b7a4-9ee12dc0d113" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.045996 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db46e68-7be5-42fd-b7a4-9ee12dc0d113" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.046013 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="978eb551-f0e5-4448-a66a-5b8256b93312" containerName="pruner" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.046499 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.049342 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.050734 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.050952 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.212518 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.213157 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.316401 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.316533 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.316580 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.363089 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.392021 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.452904 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:10:19 crc kubenswrapper[4800]: I0122 00:10:19.452978 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.045290 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hgwx4" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.045471 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-sxp25" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.159543 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.159846 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2x7st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cb9ln_openshift-marketplace(e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.161163 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cb9ln" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.203461 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.204840 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vrn8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2gn8v_openshift-marketplace(0ff9facc-6840-481a-80d2-c698c359564c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:21 crc kubenswrapper[4800]: E0122 00:10:21.206276 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2gn8v" podUID="0ff9facc-6840-481a-80d2-c698c359564c" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.425031 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.426253 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.434431 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.475921 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.476258 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.476442 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.577229 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.577308 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.577334 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.577422 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.577440 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.595499 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access\") pod \"installer-9-crc\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:23 crc kubenswrapper[4800]: I0122 00:10:23.748743 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:10:25 crc kubenswrapper[4800]: E0122 00:10:25.579096 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2gn8v" podUID="0ff9facc-6840-481a-80d2-c698c359564c" Jan 22 00:10:25 crc kubenswrapper[4800]: E0122 00:10:25.579566 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cb9ln" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.697727 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.705357 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca\") pod \"d6adcf93-e68c-4766-85f9-948c65a0d397\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.705498 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wd2f\" (UniqueName: \"kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f\") pod \"d6adcf93-e68c-4766-85f9-948c65a0d397\" (UID: \"d6adcf93-e68c-4766-85f9-948c65a0d397\") " Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.707424 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca" (OuterVolumeSpecName: "serviceca") pod "d6adcf93-e68c-4766-85f9-948c65a0d397" (UID: "d6adcf93-e68c-4766-85f9-948c65a0d397"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.734201 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f" (OuterVolumeSpecName: "kube-api-access-4wd2f") pod "d6adcf93-e68c-4766-85f9-948c65a0d397" (UID: "d6adcf93-e68c-4766-85f9-948c65a0d397"). InnerVolumeSpecName "kube-api-access-4wd2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.738671 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29484000-kr6nl" event={"ID":"d6adcf93-e68c-4766-85f9-948c65a0d397","Type":"ContainerDied","Data":"c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9"} Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.738752 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29484000-kr6nl" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.738761 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c26b5f12caf44b9ea3113c23cbf7f55047ffc79dad3fdad0a926ec41839365e9" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.806330 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wd2f\" (UniqueName: \"kubernetes.io/projected/d6adcf93-e68c-4766-85f9-948c65a0d397-kube-api-access-4wd2f\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:25 crc kubenswrapper[4800]: I0122 00:10:25.806374 4800 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d6adcf93-e68c-4766-85f9-948c65a0d397-serviceca\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:26 crc kubenswrapper[4800]: I0122 00:10:26.085903 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 00:10:26 crc kubenswrapper[4800]: W0122 00:10:26.092697 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5dfdfa65_9944_4064_a852_b6bc9833ffa8.slice/crio-b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b WatchSource:0}: Error finding container b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b: Status 404 returned error can't find the container with id b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b Jan 22 00:10:26 crc kubenswrapper[4800]: I0122 00:10:26.107425 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 00:10:26 crc kubenswrapper[4800]: I0122 00:10:26.755483 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5dfdfa65-9944-4064-a852-b6bc9833ffa8","Type":"ContainerStarted","Data":"b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b"} Jan 22 00:10:26 crc kubenswrapper[4800]: I0122 00:10:26.763415 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c4cb8420-640e-467e-b121-1e29959d187f","Type":"ContainerStarted","Data":"0cc9a1487d22ce86feccd4a5b9e39d8c8a1afa03a19033101fd89c232a88d872"} Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.817785 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.818286 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kbcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jpdkg_openshift-marketplace(859ef7fc-5aa4-4db6-bf78-e39281cbe21c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.819396 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jpdkg" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.887477 4800 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.887696 4800 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tgcj5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xdbc4_openshift-marketplace(235a6d0b-2451-4654-8cd7-b9e57ff33c4f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 00:10:26 crc kubenswrapper[4800]: E0122 00:10:26.889165 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xdbc4" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.771259 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerStarted","Data":"4097892c347b1a16f54b931e7608c227ddfd0a2ab4c1234249b92b2eb161b1f7"} Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.772789 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svkb9" event={"ID":"ca19c294-fff4-4fe6-a1a4-25e04d0624b3","Type":"ContainerStarted","Data":"0e43c9743affb0251f57104478b89c39da2f0f2946a29aa1a0a929c9c4bb83e2"} Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.775762 4800 generic.go:334] "Generic (PLEG): container finished" podID="c4cb8420-640e-467e-b121-1e29959d187f" containerID="e94495df0fd509b1d6c4fa27a9309d21a2f468658e280ce00bff37ce70e98a8a" exitCode=0 Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.775812 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c4cb8420-640e-467e-b121-1e29959d187f","Type":"ContainerDied","Data":"e94495df0fd509b1d6c4fa27a9309d21a2f468658e280ce00bff37ce70e98a8a"} Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.778027 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5dfdfa65-9944-4064-a852-b6bc9833ffa8","Type":"ContainerStarted","Data":"30fbca53cfb0e130f774103597b5e5b3452d5487268655fd04b13a7546bb70f5"} Jan 22 00:10:27 crc kubenswrapper[4800]: E0122 00:10:27.782123 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xdbc4" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" Jan 22 00:10:27 crc kubenswrapper[4800]: E0122 00:10:27.782186 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jpdkg" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.872296 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-svkb9" podStartSLOduration=185.872271357 podStartE2EDuration="3m5.872271357s" podCreationTimestamp="2026-01-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:10:27.867725325 +0000 UTC m=+203.348028363" watchObservedRunningTime="2026-01-22 00:10:27.872271357 +0000 UTC m=+203.352574405" Jan 22 00:10:27 crc kubenswrapper[4800]: I0122 00:10:27.890526 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.890506675 podStartE2EDuration="4.890506675s" podCreationTimestamp="2026-01-22 00:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:10:27.887263958 +0000 UTC m=+203.367566996" watchObservedRunningTime="2026-01-22 00:10:27.890506675 +0000 UTC m=+203.370809713" Jan 22 00:10:28 crc kubenswrapper[4800]: I0122 00:10:28.796611 4800 generic.go:334] "Generic (PLEG): container finished" podID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerID="8748a52f9f3d3c4e522463c4d3c8f09c3ce7dcd196629a422cbcf2b9b24bab05" exitCode=0 Jan 22 00:10:28 crc kubenswrapper[4800]: I0122 00:10:28.796697 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerDied","Data":"8748a52f9f3d3c4e522463c4d3c8f09c3ce7dcd196629a422cbcf2b9b24bab05"} Jan 22 00:10:28 crc kubenswrapper[4800]: I0122 00:10:28.800413 4800 generic.go:334] "Generic (PLEG): container finished" podID="6935c82f-f42c-44c6-90c5-968f24feed80" containerID="4097892c347b1a16f54b931e7608c227ddfd0a2ab4c1234249b92b2eb161b1f7" exitCode=0 Jan 22 00:10:28 crc kubenswrapper[4800]: I0122 00:10:28.800854 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerDied","Data":"4097892c347b1a16f54b931e7608c227ddfd0a2ab4c1234249b92b2eb161b1f7"} Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.066635 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.257204 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir\") pod \"c4cb8420-640e-467e-b121-1e29959d187f\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.257294 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access\") pod \"c4cb8420-640e-467e-b121-1e29959d187f\" (UID: \"c4cb8420-640e-467e-b121-1e29959d187f\") " Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.257989 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c4cb8420-640e-467e-b121-1e29959d187f" (UID: "c4cb8420-640e-467e-b121-1e29959d187f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.269115 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c4cb8420-640e-467e-b121-1e29959d187f" (UID: "c4cb8420-640e-467e-b121-1e29959d187f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.360016 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4cb8420-640e-467e-b121-1e29959d187f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.360560 4800 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c4cb8420-640e-467e-b121-1e29959d187f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.808124 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerStarted","Data":"ec881742b6943307c271c681c9e6725d1cbff1daa6f7a8c7dd5a631594a1ca89"} Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.812324 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerStarted","Data":"be547d3d5be8b049f8be3ec27e0a11f1104e6cd598f036d46370a318725a99eb"} Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.815505 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c4cb8420-640e-467e-b121-1e29959d187f","Type":"ContainerDied","Data":"0cc9a1487d22ce86feccd4a5b9e39d8c8a1afa03a19033101fd89c232a88d872"} Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.815545 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cc9a1487d22ce86feccd4a5b9e39d8c8a1afa03a19033101fd89c232a88d872" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.815526 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.833704 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-577ft" podStartSLOduration=2.73737437 podStartE2EDuration="52.833687949s" podCreationTimestamp="2026-01-22 00:09:37 +0000 UTC" firstStartedPulling="2026-01-22 00:09:39.183364226 +0000 UTC m=+154.663667264" lastFinishedPulling="2026-01-22 00:10:29.279677805 +0000 UTC m=+204.759980843" observedRunningTime="2026-01-22 00:10:29.829663301 +0000 UTC m=+205.309966339" watchObservedRunningTime="2026-01-22 00:10:29.833687949 +0000 UTC m=+205.313990987" Jan 22 00:10:29 crc kubenswrapper[4800]: I0122 00:10:29.851081 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n28k4" podStartSLOduration=3.82694764 podStartE2EDuration="53.851059914s" podCreationTimestamp="2026-01-22 00:09:36 +0000 UTC" firstStartedPulling="2026-01-22 00:09:39.199745758 +0000 UTC m=+154.680048796" lastFinishedPulling="2026-01-22 00:10:29.223858032 +0000 UTC m=+204.704161070" observedRunningTime="2026-01-22 00:10:29.849525733 +0000 UTC m=+205.329828781" watchObservedRunningTime="2026-01-22 00:10:29.851059914 +0000 UTC m=+205.331362952" Jan 22 00:10:36 crc kubenswrapper[4800]: I0122 00:10:36.874926 4800 generic.go:334] "Generic (PLEG): container finished" podID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerID="8a22ea244137470f029ea0968141959c029459ba2602cbe3a0d33da8866ab207" exitCode=0 Jan 22 00:10:36 crc kubenswrapper[4800]: I0122 00:10:36.875655 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerDied","Data":"8a22ea244137470f029ea0968141959c029459ba2602cbe3a0d33da8866ab207"} Jan 22 00:10:36 crc kubenswrapper[4800]: I0122 00:10:36.883811 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerID="1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7" exitCode=0 Jan 22 00:10:36 crc kubenswrapper[4800]: I0122 00:10:36.883854 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerDied","Data":"1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7"} Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.112920 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.113051 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.187638 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.481827 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.482002 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.526708 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.891789 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerStarted","Data":"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a"} Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.895797 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerStarted","Data":"b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185"} Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.944071 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sxp25" podStartSLOduration=3.6780152 podStartE2EDuration="1m3.944045164s" podCreationTimestamp="2026-01-22 00:09:34 +0000 UTC" firstStartedPulling="2026-01-22 00:09:37.010655013 +0000 UTC m=+152.490958051" lastFinishedPulling="2026-01-22 00:10:37.276684977 +0000 UTC m=+212.756988015" observedRunningTime="2026-01-22 00:10:37.920990787 +0000 UTC m=+213.401293825" watchObservedRunningTime="2026-01-22 00:10:37.944045164 +0000 UTC m=+213.424348202" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.954839 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.963380 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:37 crc kubenswrapper[4800]: I0122 00:10:37.987034 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hgwx4" podStartSLOduration=4.444664888 podStartE2EDuration="1m4.987009114s" podCreationTimestamp="2026-01-22 00:09:33 +0000 UTC" firstStartedPulling="2026-01-22 00:09:36.794254795 +0000 UTC m=+152.274557833" lastFinishedPulling="2026-01-22 00:10:37.336599021 +0000 UTC m=+212.816902059" observedRunningTime="2026-01-22 00:10:37.947090666 +0000 UTC m=+213.427393704" watchObservedRunningTime="2026-01-22 00:10:37.987009114 +0000 UTC m=+213.467312142" Jan 22 00:10:39 crc kubenswrapper[4800]: I0122 00:10:39.347352 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:10:40 crc kubenswrapper[4800]: I0122 00:10:40.918747 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-577ft" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="registry-server" containerID="cri-o://ec881742b6943307c271c681c9e6725d1cbff1daa6f7a8c7dd5a631594a1ca89" gracePeriod=2 Jan 22 00:10:41 crc kubenswrapper[4800]: I0122 00:10:41.925717 4800 generic.go:334] "Generic (PLEG): container finished" podID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerID="ec881742b6943307c271c681c9e6725d1cbff1daa6f7a8c7dd5a631594a1ca89" exitCode=0 Jan 22 00:10:41 crc kubenswrapper[4800]: I0122 00:10:41.925809 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerDied","Data":"ec881742b6943307c271c681c9e6725d1cbff1daa6f7a8c7dd5a631594a1ca89"} Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.304298 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.304378 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.356064 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.455373 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.455435 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.502153 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.946150 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-577ft" event={"ID":"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04","Type":"ContainerDied","Data":"bfe3fdbf911965fc3397b65841b93fce985432a0caea838089f3cb0f6d3c7538"} Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.946459 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfe3fdbf911965fc3397b65841b93fce985432a0caea838089f3cb0f6d3c7538" Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.950903 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerStarted","Data":"e81920cb983bb46f701a9cb3a433c223b486a07c969f1e180b170732532c461f"} Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.957215 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerStarted","Data":"9c09c3c8069e9612a8db2124a43bc1ad85d1351ecaecad657e6bca37866a4226"} Jan 22 00:10:44 crc kubenswrapper[4800]: I0122 00:10:44.974072 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.022862 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.032278 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.079320 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content\") pod \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.079439 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities\") pod \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.080518 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities" (OuterVolumeSpecName: "utilities") pod "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" (UID: "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.080627 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw44b\" (UniqueName: \"kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b\") pod \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\" (UID: \"c8ee1ff3-44c1-4993-91e7-bb5ba795ea04\") " Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.081773 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.088143 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b" (OuterVolumeSpecName: "kube-api-access-gw44b") pod "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" (UID: "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04"). InnerVolumeSpecName "kube-api-access-gw44b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.183391 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw44b\" (UniqueName: \"kubernetes.io/projected/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-kube-api-access-gw44b\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.217972 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" (UID: "c8ee1ff3-44c1-4993-91e7-bb5ba795ea04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.284561 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.944500 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.981222 4800 generic.go:334] "Generic (PLEG): container finished" podID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerID="b6a1b1fb2154816859c0dd1cd03b25c3842ef1ced846dc59af991d1f59a1b653" exitCode=0 Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.981428 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerDied","Data":"b6a1b1fb2154816859c0dd1cd03b25c3842ef1ced846dc59af991d1f59a1b653"} Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.982829 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ff9facc-6840-481a-80d2-c698c359564c" containerID="9c09c3c8069e9612a8db2124a43bc1ad85d1351ecaecad657e6bca37866a4226" exitCode=0 Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.983004 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerDied","Data":"9c09c3c8069e9612a8db2124a43bc1ad85d1351ecaecad657e6bca37866a4226"} Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.986848 4800 generic.go:334] "Generic (PLEG): container finished" podID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerID="35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a" exitCode=0 Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.986938 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerDied","Data":"35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a"} Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.990040 4800 generic.go:334] "Generic (PLEG): container finished" podID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerID="e81920cb983bb46f701a9cb3a433c223b486a07c969f1e180b170732532c461f" exitCode=0 Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.990156 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-577ft" Jan 22 00:10:45 crc kubenswrapper[4800]: I0122 00:10:45.990633 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerDied","Data":"e81920cb983bb46f701a9cb3a433c223b486a07c969f1e180b170732532c461f"} Jan 22 00:10:46 crc kubenswrapper[4800]: I0122 00:10:46.070410 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:10:46 crc kubenswrapper[4800]: I0122 00:10:46.073028 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-577ft"] Jan 22 00:10:46 crc kubenswrapper[4800]: I0122 00:10:46.823865 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" path="/var/lib/kubelet/pods/c8ee1ff3-44c1-4993-91e7-bb5ba795ea04/volumes" Jan 22 00:10:46 crc kubenswrapper[4800]: I0122 00:10:46.998424 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerStarted","Data":"a38209bfe29bf43f5ccac55ca51c0b897cc97c1cff29223ec420e5a2db5545fb"} Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.000394 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerStarted","Data":"719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3"} Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.002251 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerStarted","Data":"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354"} Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.004178 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerStarted","Data":"c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26"} Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.004390 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sxp25" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="registry-server" containerID="cri-o://90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a" gracePeriod=2 Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.043343 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpdkg" podStartSLOduration=3.682577903 podStartE2EDuration="1m12.04332563s" podCreationTimestamp="2026-01-22 00:09:35 +0000 UTC" firstStartedPulling="2026-01-22 00:09:38.106038937 +0000 UTC m=+153.586341975" lastFinishedPulling="2026-01-22 00:10:46.466786664 +0000 UTC m=+221.947089702" observedRunningTime="2026-01-22 00:10:47.025107242 +0000 UTC m=+222.505410280" watchObservedRunningTime="2026-01-22 00:10:47.04332563 +0000 UTC m=+222.523628658" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.043826 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xdbc4" podStartSLOduration=3.774158892 podStartE2EDuration="1m12.043820214s" podCreationTimestamp="2026-01-22 00:09:35 +0000 UTC" firstStartedPulling="2026-01-22 00:09:38.14215217 +0000 UTC m=+153.622455218" lastFinishedPulling="2026-01-22 00:10:46.411813492 +0000 UTC m=+221.892116540" observedRunningTime="2026-01-22 00:10:47.041732647 +0000 UTC m=+222.522035685" watchObservedRunningTime="2026-01-22 00:10:47.043820214 +0000 UTC m=+222.524123252" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.078325 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cb9ln" podStartSLOduration=4.688697178 podStartE2EDuration="1m14.078299275s" podCreationTimestamp="2026-01-22 00:09:33 +0000 UTC" firstStartedPulling="2026-01-22 00:09:36.970557073 +0000 UTC m=+152.450860111" lastFinishedPulling="2026-01-22 00:10:46.36015917 +0000 UTC m=+221.840462208" observedRunningTime="2026-01-22 00:10:47.074480964 +0000 UTC m=+222.554784002" watchObservedRunningTime="2026-01-22 00:10:47.078299275 +0000 UTC m=+222.558602313" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.108530 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2gn8v" podStartSLOduration=4.286579638 podStartE2EDuration="1m14.108506024s" podCreationTimestamp="2026-01-22 00:09:33 +0000 UTC" firstStartedPulling="2026-01-22 00:09:36.624857712 +0000 UTC m=+152.105160750" lastFinishedPulling="2026-01-22 00:10:46.446784098 +0000 UTC m=+221.927087136" observedRunningTime="2026-01-22 00:10:47.10424934 +0000 UTC m=+222.584552378" watchObservedRunningTime="2026-01-22 00:10:47.108506024 +0000 UTC m=+222.588809062" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.480728 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.614251 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fc7g\" (UniqueName: \"kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g\") pod \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.614354 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content\") pod \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.614496 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities\") pod \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\" (UID: \"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4\") " Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.615626 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities" (OuterVolumeSpecName: "utilities") pod "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" (UID: "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.625466 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g" (OuterVolumeSpecName: "kube-api-access-9fc7g") pod "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" (UID: "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4"). InnerVolumeSpecName "kube-api-access-9fc7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.665680 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" (UID: "0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.716366 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.716425 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:47 crc kubenswrapper[4800]: I0122 00:10:47.716439 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fc7g\" (UniqueName: \"kubernetes.io/projected/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4-kube-api-access-9fc7g\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.012658 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerID="90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a" exitCode=0 Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.012729 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerDied","Data":"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a"} Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.012771 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxp25" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.012828 4800 scope.go:117] "RemoveContainer" containerID="90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.012795 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxp25" event={"ID":"0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4","Type":"ContainerDied","Data":"ea42d617f2dd2f5139f31b3c8057f8e41d573e19ef8eceb3bd915b7c77f73424"} Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.041161 4800 scope.go:117] "RemoveContainer" containerID="1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.067716 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.070447 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sxp25"] Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.074076 4800 scope.go:117] "RemoveContainer" containerID="f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.092424 4800 scope.go:117] "RemoveContainer" containerID="90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a" Jan 22 00:10:48 crc kubenswrapper[4800]: E0122 00:10:48.093069 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a\": container with ID starting with 90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a not found: ID does not exist" containerID="90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.093109 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a"} err="failed to get container status \"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a\": rpc error: code = NotFound desc = could not find container \"90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a\": container with ID starting with 90dc1f4e5246e4cf775f9ab393d3722f8363e8418f2693695550f57e582c038a not found: ID does not exist" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.093169 4800 scope.go:117] "RemoveContainer" containerID="1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7" Jan 22 00:10:48 crc kubenswrapper[4800]: E0122 00:10:48.093562 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7\": container with ID starting with 1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7 not found: ID does not exist" containerID="1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.093589 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7"} err="failed to get container status \"1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7\": rpc error: code = NotFound desc = could not find container \"1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7\": container with ID starting with 1847857160d594d8ce36cd902895ed2df45f5142821a846633e138e89c0e76d7 not found: ID does not exist" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.093632 4800 scope.go:117] "RemoveContainer" containerID="f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf" Jan 22 00:10:48 crc kubenswrapper[4800]: E0122 00:10:48.094365 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf\": container with ID starting with f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf not found: ID does not exist" containerID="f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.094393 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf"} err="failed to get container status \"f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf\": rpc error: code = NotFound desc = could not find container \"f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf\": container with ID starting with f342e3fdb783907c41c700a52c32a8a943fa6471992603067ca471fc3add64cf not found: ID does not exist" Jan 22 00:10:48 crc kubenswrapper[4800]: I0122 00:10:48.827219 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" path="/var/lib/kubelet/pods/0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4/volumes" Jan 22 00:10:49 crc kubenswrapper[4800]: I0122 00:10:49.451542 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:10:49 crc kubenswrapper[4800]: I0122 00:10:49.452013 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:10:49 crc kubenswrapper[4800]: I0122 00:10:49.452076 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:10:49 crc kubenswrapper[4800]: I0122 00:10:49.452782 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:10:49 crc kubenswrapper[4800]: I0122 00:10:49.452838 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7" gracePeriod=600 Jan 22 00:10:50 crc kubenswrapper[4800]: I0122 00:10:50.030277 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7" exitCode=0 Jan 22 00:10:50 crc kubenswrapper[4800]: I0122 00:10:50.030331 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7"} Jan 22 00:10:50 crc kubenswrapper[4800]: I0122 00:10:50.030924 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976"} Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.045339 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.046038 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.088186 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.136564 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.347230 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.347276 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:10:54 crc kubenswrapper[4800]: I0122 00:10:54.388399 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:10:55 crc kubenswrapper[4800]: I0122 00:10:55.107396 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.039332 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.039767 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.101411 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.150218 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.371271 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.373315 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.417227 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:10:56 crc kubenswrapper[4800]: I0122 00:10:56.907421 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq4b8"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.139028 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.299516 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.300208 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hgwx4" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="registry-server" containerID="cri-o://b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185" gracePeriod=30 Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.315359 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.315657 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2gn8v" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="registry-server" containerID="cri-o://719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3" gracePeriod=30 Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.324926 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.325177 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cb9ln" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="registry-server" containerID="cri-o://c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26" gracePeriod=30 Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.346371 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.346582 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" containerID="cri-o://ae93bae088e74583820f91b8ff08f1b07e5567e6a65e46b7ef4c836c367bc392" gracePeriod=30 Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.360166 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.372376 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375369 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8d98k"] Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375583 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="extract-content" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375599 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="extract-content" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375610 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375617 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375628 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cb8420-640e-467e-b121-1e29959d187f" containerName="pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375634 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cb8420-640e-467e-b121-1e29959d187f" containerName="pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375643 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="extract-utilities" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375649 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="extract-utilities" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375659 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="extract-utilities" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375665 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="extract-utilities" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375674 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="extract-content" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375679 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="extract-content" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375688 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375695 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.375707 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6adcf93-e68c-4766-85f9-948c65a0d397" containerName="image-pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375714 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6adcf93-e68c-4766-85f9-948c65a0d397" containerName="image-pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375797 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6adcf93-e68c-4766-85f9-948c65a0d397" containerName="image-pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375808 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ebf236f-c8eb-40f6-8bf9-9d3e092b97c4" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375821 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cb8420-640e-467e-b121-1e29959d187f" containerName="pruner" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.375828 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ee1ff3-44c1-4993-91e7-bb5ba795ea04" containerName="registry-server" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.376195 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.384970 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.385299 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n28k4" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="registry-server" containerID="cri-o://be547d3d5be8b049f8be3ec27e0a11f1104e6cd598f036d46370a318725a99eb" gracePeriod=30 Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.430094 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8d98k"] Jan 22 00:10:57 crc kubenswrapper[4800]: E0122 00:10:57.434733 4800 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19d5c725_0286_404e_9f66_a22ed5c2649c.slice/crio-b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4bf8eaf_1086_4fdc_9327_bb2fb26ad7e6.slice/crio-c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ff9facc_6840_481a_80d2_c698c359564c.slice/crio-719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3.scope\": RecentStats: unable to find data in memory cache]" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.451023 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv8jb\" (UniqueName: \"kubernetes.io/projected/a816e793-2455-4446-85dd-6437bc1d6ee7-kube-api-access-zv8jb\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.451160 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.451196 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.552383 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv8jb\" (UniqueName: \"kubernetes.io/projected/a816e793-2455-4446-85dd-6437bc1d6ee7-kube-api-access-zv8jb\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.552454 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.552488 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.554444 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.559573 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a816e793-2455-4446-85dd-6437bc1d6ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.574253 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv8jb\" (UniqueName: \"kubernetes.io/projected/a816e793-2455-4446-85dd-6437bc1d6ee7-kube-api-access-zv8jb\") pod \"marketplace-operator-79b997595-8d98k\" (UID: \"a816e793-2455-4446-85dd-6437bc1d6ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:57 crc kubenswrapper[4800]: I0122 00:10:57.703571 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:58 crc kubenswrapper[4800]: I0122 00:10:58.087006 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jpdkg" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="registry-server" containerID="cri-o://a38209bfe29bf43f5ccac55ca51c0b897cc97c1cff29223ec420e5a2db5545fb" gracePeriod=30 Jan 22 00:10:58 crc kubenswrapper[4800]: I0122 00:10:58.095308 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8d98k"] Jan 22 00:10:58 crc kubenswrapper[4800]: W0122 00:10:58.110810 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda816e793_2455_4446_85dd_6437bc1d6ee7.slice/crio-d1f5ee1dfc7415c74f4f49d9f07a17491901b36f0e6a159fcb84692e574ece62 WatchSource:0}: Error finding container d1f5ee1dfc7415c74f4f49d9f07a17491901b36f0e6a159fcb84692e574ece62: Status 404 returned error can't find the container with id d1f5ee1dfc7415c74f4f49d9f07a17491901b36f0e6a159fcb84692e574ece62 Jan 22 00:10:58 crc kubenswrapper[4800]: I0122 00:10:58.343292 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:10:58 crc kubenswrapper[4800]: I0122 00:10:58.898601 4800 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r9z5h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Jan 22 00:10:58 crc kubenswrapper[4800]: I0122 00:10:58.898663 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.094565 4800 generic.go:334] "Generic (PLEG): container finished" podID="0ff9facc-6840-481a-80d2-c698c359564c" containerID="719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.094657 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerDied","Data":"719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.096311 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" event={"ID":"a816e793-2455-4446-85dd-6437bc1d6ee7","Type":"ContainerStarted","Data":"49abbc92e5b4b201c4328e70b8cd7b89f86a0d2768a1022c7ab5528f26ab90d6"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.096349 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" event={"ID":"a816e793-2455-4446-85dd-6437bc1d6ee7","Type":"ContainerStarted","Data":"d1f5ee1dfc7415c74f4f49d9f07a17491901b36f0e6a159fcb84692e574ece62"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.096465 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.098676 4800 generic.go:334] "Generic (PLEG): container finished" podID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerID="ae93bae088e74583820f91b8ff08f1b07e5567e6a65e46b7ef4c836c367bc392" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.098747 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" event={"ID":"7446e4e9-c6b6-48e5-8488-6484f9f274ae","Type":"ContainerDied","Data":"ae93bae088e74583820f91b8ff08f1b07e5567e6a65e46b7ef4c836c367bc392"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.099133 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.103296 4800 generic.go:334] "Generic (PLEG): container finished" podID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerID="b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.103361 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerDied","Data":"b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.105662 4800 generic.go:334] "Generic (PLEG): container finished" podID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerID="c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.105690 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerDied","Data":"c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.108301 4800 generic.go:334] "Generic (PLEG): container finished" podID="6935c82f-f42c-44c6-90c5-968f24feed80" containerID="be547d3d5be8b049f8be3ec27e0a11f1104e6cd598f036d46370a318725a99eb" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.108373 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerDied","Data":"be547d3d5be8b049f8be3ec27e0a11f1104e6cd598f036d46370a318725a99eb"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.111787 4800 generic.go:334] "Generic (PLEG): container finished" podID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerID="a38209bfe29bf43f5ccac55ca51c0b897cc97c1cff29223ec420e5a2db5545fb" exitCode=0 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.111864 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerDied","Data":"a38209bfe29bf43f5ccac55ca51c0b897cc97c1cff29223ec420e5a2db5545fb"} Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.112018 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xdbc4" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="registry-server" containerID="cri-o://dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354" gracePeriod=30 Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.113046 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8d98k" podStartSLOduration=2.113034438 podStartE2EDuration="2.113034438s" podCreationTimestamp="2026-01-22 00:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:10:59.112033711 +0000 UTC m=+234.592336749" watchObservedRunningTime="2026-01-22 00:10:59.113034438 +0000 UTC m=+234.593337476" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.349155 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.440500 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.573405 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities\") pod \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.573466 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kbcg\" (UniqueName: \"kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg\") pod \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.573533 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content\") pod \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\" (UID: \"859ef7fc-5aa4-4db6-bf78-e39281cbe21c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.574910 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities" (OuterVolumeSpecName: "utilities") pod "859ef7fc-5aa4-4db6-bf78-e39281cbe21c" (UID: "859ef7fc-5aa4-4db6-bf78-e39281cbe21c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.585343 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg" (OuterVolumeSpecName: "kube-api-access-6kbcg") pod "859ef7fc-5aa4-4db6-bf78-e39281cbe21c" (UID: "859ef7fc-5aa4-4db6-bf78-e39281cbe21c"). InnerVolumeSpecName "kube-api-access-6kbcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.600708 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "859ef7fc-5aa4-4db6-bf78-e39281cbe21c" (UID: "859ef7fc-5aa4-4db6-bf78-e39281cbe21c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.614830 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.674678 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content\") pod \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.674754 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities\") pod \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.674810 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgcj5\" (UniqueName: \"kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5\") pod \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\" (UID: \"235a6d0b-2451-4654-8cd7-b9e57ff33c4f\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.675052 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.675067 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kbcg\" (UniqueName: \"kubernetes.io/projected/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-kube-api-access-6kbcg\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.675078 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859ef7fc-5aa4-4db6-bf78-e39281cbe21c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.675578 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities" (OuterVolumeSpecName: "utilities") pod "235a6d0b-2451-4654-8cd7-b9e57ff33c4f" (UID: "235a6d0b-2451-4654-8cd7-b9e57ff33c4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.684230 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5" (OuterVolumeSpecName: "kube-api-access-tgcj5") pod "235a6d0b-2451-4654-8cd7-b9e57ff33c4f" (UID: "235a6d0b-2451-4654-8cd7-b9e57ff33c4f"). InnerVolumeSpecName "kube-api-access-tgcj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.700285 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "235a6d0b-2451-4654-8cd7-b9e57ff33c4f" (UID: "235a6d0b-2451-4654-8cd7-b9e57ff33c4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.748004 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.775939 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgcj5\" (UniqueName: \"kubernetes.io/projected/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-kube-api-access-tgcj5\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.775974 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.775988 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/235a6d0b-2451-4654-8cd7-b9e57ff33c4f-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.800021 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.865604 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876711 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities\") pod \"19d5c725-0286-404e-9f66-a22ed5c2649c\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876766 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content\") pod \"19d5c725-0286-404e-9f66-a22ed5c2649c\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876811 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics\") pod \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876841 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvffz\" (UniqueName: \"kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz\") pod \"19d5c725-0286-404e-9f66-a22ed5c2649c\" (UID: \"19d5c725-0286-404e-9f66-a22ed5c2649c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876952 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca\") pod \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.876997 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kswd\" (UniqueName: \"kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd\") pod \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\" (UID: \"7446e4e9-c6b6-48e5-8488-6484f9f274ae\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.877481 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities" (OuterVolumeSpecName: "utilities") pod "19d5c725-0286-404e-9f66-a22ed5c2649c" (UID: "19d5c725-0286-404e-9f66-a22ed5c2649c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.879071 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7446e4e9-c6b6-48e5-8488-6484f9f274ae" (UID: "7446e4e9-c6b6-48e5-8488-6484f9f274ae"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.881245 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz" (OuterVolumeSpecName: "kube-api-access-rvffz") pod "19d5c725-0286-404e-9f66-a22ed5c2649c" (UID: "19d5c725-0286-404e-9f66-a22ed5c2649c"). InnerVolumeSpecName "kube-api-access-rvffz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.882375 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7446e4e9-c6b6-48e5-8488-6484f9f274ae" (UID: "7446e4e9-c6b6-48e5-8488-6484f9f274ae"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.884260 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd" (OuterVolumeSpecName: "kube-api-access-8kswd") pod "7446e4e9-c6b6-48e5-8488-6484f9f274ae" (UID: "7446e4e9-c6b6-48e5-8488-6484f9f274ae"). InnerVolumeSpecName "kube-api-access-8kswd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.889690 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.893291 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.960768 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19d5c725-0286-404e-9f66-a22ed5c2649c" (UID: "19d5c725-0286-404e-9f66-a22ed5c2649c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978056 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content\") pod \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978111 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk42q\" (UniqueName: \"kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q\") pod \"6935c82f-f42c-44c6-90c5-968f24feed80\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978169 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content\") pod \"6935c82f-f42c-44c6-90c5-968f24feed80\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978194 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x7st\" (UniqueName: \"kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st\") pod \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978223 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities\") pod \"0ff9facc-6840-481a-80d2-c698c359564c\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978239 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities\") pod \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\" (UID: \"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978256 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrn8x\" (UniqueName: \"kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x\") pod \"0ff9facc-6840-481a-80d2-c698c359564c\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978272 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content\") pod \"0ff9facc-6840-481a-80d2-c698c359564c\" (UID: \"0ff9facc-6840-481a-80d2-c698c359564c\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978308 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities\") pod \"6935c82f-f42c-44c6-90c5-968f24feed80\" (UID: \"6935c82f-f42c-44c6-90c5-968f24feed80\") " Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978501 4800 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978515 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kswd\" (UniqueName: \"kubernetes.io/projected/7446e4e9-c6b6-48e5-8488-6484f9f274ae-kube-api-access-8kswd\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978525 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978532 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d5c725-0286-404e-9f66-a22ed5c2649c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978540 4800 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7446e4e9-c6b6-48e5-8488-6484f9f274ae-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.978555 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvffz\" (UniqueName: \"kubernetes.io/projected/19d5c725-0286-404e-9f66-a22ed5c2649c-kube-api-access-rvffz\") on node \"crc\" DevicePath \"\"" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.979195 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities" (OuterVolumeSpecName: "utilities") pod "6935c82f-f42c-44c6-90c5-968f24feed80" (UID: "6935c82f-f42c-44c6-90c5-968f24feed80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.980410 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities" (OuterVolumeSpecName: "utilities") pod "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" (UID: "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.981873 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities" (OuterVolumeSpecName: "utilities") pod "0ff9facc-6840-481a-80d2-c698c359564c" (UID: "0ff9facc-6840-481a-80d2-c698c359564c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.982951 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x" (OuterVolumeSpecName: "kube-api-access-vrn8x") pod "0ff9facc-6840-481a-80d2-c698c359564c" (UID: "0ff9facc-6840-481a-80d2-c698c359564c"). InnerVolumeSpecName "kube-api-access-vrn8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.983021 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q" (OuterVolumeSpecName: "kube-api-access-xk42q") pod "6935c82f-f42c-44c6-90c5-968f24feed80" (UID: "6935c82f-f42c-44c6-90c5-968f24feed80"). InnerVolumeSpecName "kube-api-access-xk42q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:10:59 crc kubenswrapper[4800]: I0122 00:10:59.983501 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st" (OuterVolumeSpecName: "kube-api-access-2x7st") pod "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" (UID: "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6"). InnerVolumeSpecName "kube-api-access-2x7st". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.048813 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ff9facc-6840-481a-80d2-c698c359564c" (UID: "0ff9facc-6840-481a-80d2-c698c359564c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.054104 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" (UID: "e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079600 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079647 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079660 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrn8x\" (UniqueName: \"kubernetes.io/projected/0ff9facc-6840-481a-80d2-c698c359564c-kube-api-access-vrn8x\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079672 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9facc-6840-481a-80d2-c698c359564c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079683 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079695 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079708 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk42q\" (UniqueName: \"kubernetes.io/projected/6935c82f-f42c-44c6-90c5-968f24feed80-kube-api-access-xk42q\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.079720 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x7st\" (UniqueName: \"kubernetes.io/projected/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6-kube-api-access-2x7st\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.113137 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6935c82f-f42c-44c6-90c5-968f24feed80" (UID: "6935c82f-f42c-44c6-90c5-968f24feed80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.119534 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb9ln" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.119544 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb9ln" event={"ID":"e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6","Type":"ContainerDied","Data":"b430618eda7d5a2619933a352aef6706b7ac948a1e5473c18d4f988b1b4b659b"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.119603 4800 scope.go:117] "RemoveContainer" containerID="c69f0516ffb5ceccb9e06080108b8df90afcbe2b8f5d36e6fd26780c1104ea26" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.124458 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n28k4" event={"ID":"6935c82f-f42c-44c6-90c5-968f24feed80","Type":"ContainerDied","Data":"13ea5a7aceb71b131dd37e2acc87eb726edc23fc62d3073be7207d8bd22efaef"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.124473 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n28k4" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.127138 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpdkg" event={"ID":"859ef7fc-5aa4-4db6-bf78-e39281cbe21c","Type":"ContainerDied","Data":"7a24fdd8de936c9f24516bcdee24b159ae341d265ad234eb7fe6789107edf27b"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.127188 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpdkg" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.131226 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2gn8v" event={"ID":"0ff9facc-6840-481a-80d2-c698c359564c","Type":"ContainerDied","Data":"6b2bd2d3c8488e6534ed07035f49dc23546ab0da847a6a07739985da0d1cf14c"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.131334 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2gn8v" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.139016 4800 generic.go:334] "Generic (PLEG): container finished" podID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerID="dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354" exitCode=0 Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.139081 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdbc4" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.139094 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerDied","Data":"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.139128 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdbc4" event={"ID":"235a6d0b-2451-4654-8cd7-b9e57ff33c4f","Type":"ContainerDied","Data":"4cf6dc30ccdcfc5f1b177941a7b4d15e4c2015e530871bdaffde926d566d8367"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.139513 4800 scope.go:117] "RemoveContainer" containerID="e81920cb983bb46f701a9cb3a433c223b486a07c969f1e180b170732532c461f" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.144852 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" event={"ID":"7446e4e9-c6b6-48e5-8488-6484f9f274ae","Type":"ContainerDied","Data":"c20780e188eb3932d13b109f57b220106a94d787c436caa3bce067105cf66ed7"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.145001 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r9z5h" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.150758 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.151735 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgwx4" event={"ID":"19d5c725-0286-404e-9f66-a22ed5c2649c","Type":"ContainerDied","Data":"e4d666f683dc23e9f8a4bdeb3e12804863b3114690e109ec9a08a40db28997d5"} Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.151797 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgwx4" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.154190 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cb9ln"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.180012 4800 scope.go:117] "RemoveContainer" containerID="1b6b25536c8c149379b7c82a1a66bc968a2f4d2fe2bb64cce5e0d8634afc426b" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.180764 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6935c82f-f42c-44c6-90c5-968f24feed80-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.201545 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.206611 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n28k4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.209115 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.211424 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2gn8v"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.211770 4800 scope.go:117] "RemoveContainer" containerID="be547d3d5be8b049f8be3ec27e0a11f1104e6cd598f036d46370a318725a99eb" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.222945 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.225740 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdbc4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.235764 4800 scope.go:117] "RemoveContainer" containerID="4097892c347b1a16f54b931e7608c227ddfd0a2ab4c1234249b92b2eb161b1f7" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.237234 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.242250 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpdkg"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.253624 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.260176 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r9z5h"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.264694 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.267657 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hgwx4"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.272757 4800 scope.go:117] "RemoveContainer" containerID="f2ac286b7cca96fa19b30e06f29b2e412f39b4143f4a2217cefa65950d4d56ea" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.286297 4800 scope.go:117] "RemoveContainer" containerID="a38209bfe29bf43f5ccac55ca51c0b897cc97c1cff29223ec420e5a2db5545fb" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.303236 4800 scope.go:117] "RemoveContainer" containerID="b6a1b1fb2154816859c0dd1cd03b25c3842ef1ced846dc59af991d1f59a1b653" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.317822 4800 scope.go:117] "RemoveContainer" containerID="67a397d4096cc85de2f875409edf4b3335c224853dd6455f126f3fc9af7f9135" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.330908 4800 scope.go:117] "RemoveContainer" containerID="719546fb70f9e209749055268599a716faffa7c6094b39340b323b5f761b89c3" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.346384 4800 scope.go:117] "RemoveContainer" containerID="9c09c3c8069e9612a8db2124a43bc1ad85d1351ecaecad657e6bca37866a4226" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.362586 4800 scope.go:117] "RemoveContainer" containerID="8bc90665985f86ee2cba1121e3c10263afb6c10e477e9c227b206a0c659160ad" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.373559 4800 scope.go:117] "RemoveContainer" containerID="dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.385331 4800 scope.go:117] "RemoveContainer" containerID="35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.398193 4800 scope.go:117] "RemoveContainer" containerID="081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.413137 4800 scope.go:117] "RemoveContainer" containerID="dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.413522 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354\": container with ID starting with dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354 not found: ID does not exist" containerID="dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.413556 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354"} err="failed to get container status \"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354\": rpc error: code = NotFound desc = could not find container \"dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354\": container with ID starting with dba312af7d5a55d5c32992521d316b91c955f6ae99bc396fd45a09fe10990354 not found: ID does not exist" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.413584 4800 scope.go:117] "RemoveContainer" containerID="35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.413984 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a\": container with ID starting with 35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a not found: ID does not exist" containerID="35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.414005 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a"} err="failed to get container status \"35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a\": rpc error: code = NotFound desc = could not find container \"35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a\": container with ID starting with 35221b848323ac6cd1d34f725e524d61e61a9eddc324af87afb7589bae5a533a not found: ID does not exist" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.414019 4800 scope.go:117] "RemoveContainer" containerID="081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.414336 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745\": container with ID starting with 081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745 not found: ID does not exist" containerID="081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.414358 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745"} err="failed to get container status \"081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745\": rpc error: code = NotFound desc = could not find container \"081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745\": container with ID starting with 081a455977ea0d050472f75ad54f4397a5a8a7ef2ad01cfab68fe023bdb4c745 not found: ID does not exist" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.414372 4800 scope.go:117] "RemoveContainer" containerID="ae93bae088e74583820f91b8ff08f1b07e5567e6a65e46b7ef4c836c367bc392" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.427311 4800 scope.go:117] "RemoveContainer" containerID="b04a5d0aa322b96e53728e50bb9fdaff7fb89068b3123687162e72a355805185" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.442692 4800 scope.go:117] "RemoveContainer" containerID="8a22ea244137470f029ea0968141959c029459ba2602cbe3a0d33da8866ab207" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.458597 4800 scope.go:117] "RemoveContainer" containerID="8742b7bcf620717d57cd5b314c0cc37cd7fd28c3bc9a4f1d6b1a8cfef674fc39" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.750863 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x8jld"] Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751165 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751184 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751198 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751206 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751222 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751231 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751242 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751249 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751259 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751266 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751277 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751284 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751353 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751362 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751372 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751379 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751390 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751398 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751407 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751415 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751425 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751431 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751442 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751449 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="extract-content" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751460 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751467 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751540 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751548 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751560 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751568 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751576 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751584 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751594 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751602 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751611 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751619 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="extract-utilities" Jan 22 00:11:00 crc kubenswrapper[4800]: E0122 00:11:00.751629 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751636 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751750 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751765 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751775 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751785 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff9facc-6840-481a-80d2-c698c359564c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751795 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" containerName="marketplace-operator" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751810 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.751818 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" containerName="registry-server" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.752821 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.754705 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.760458 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8jld"] Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.790019 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-catalog-content\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.790068 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q5cx\" (UniqueName: \"kubernetes.io/projected/ecdd722c-78ea-4e3e-bd54-31e53b1177da-kube-api-access-2q5cx\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.790118 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-utilities\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.824014 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff9facc-6840-481a-80d2-c698c359564c" path="/var/lib/kubelet/pods/0ff9facc-6840-481a-80d2-c698c359564c/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.825017 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d5c725-0286-404e-9f66-a22ed5c2649c" path="/var/lib/kubelet/pods/19d5c725-0286-404e-9f66-a22ed5c2649c/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.825617 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235a6d0b-2451-4654-8cd7-b9e57ff33c4f" path="/var/lib/kubelet/pods/235a6d0b-2451-4654-8cd7-b9e57ff33c4f/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.827000 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6935c82f-f42c-44c6-90c5-968f24feed80" path="/var/lib/kubelet/pods/6935c82f-f42c-44c6-90c5-968f24feed80/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.827664 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7446e4e9-c6b6-48e5-8488-6484f9f274ae" path="/var/lib/kubelet/pods/7446e4e9-c6b6-48e5-8488-6484f9f274ae/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.828624 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859ef7fc-5aa4-4db6-bf78-e39281cbe21c" path="/var/lib/kubelet/pods/859ef7fc-5aa4-4db6-bf78-e39281cbe21c/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.829343 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6" path="/var/lib/kubelet/pods/e4bf8eaf-1086-4fdc-9327-bb2fb26ad7e6/volumes" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.891548 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-utilities\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.891632 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-catalog-content\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.891657 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q5cx\" (UniqueName: \"kubernetes.io/projected/ecdd722c-78ea-4e3e-bd54-31e53b1177da-kube-api-access-2q5cx\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.892056 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-utilities\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.892265 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecdd722c-78ea-4e3e-bd54-31e53b1177da-catalog-content\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:00 crc kubenswrapper[4800]: I0122 00:11:00.912739 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q5cx\" (UniqueName: \"kubernetes.io/projected/ecdd722c-78ea-4e3e-bd54-31e53b1177da-kube-api-access-2q5cx\") pod \"redhat-operators-x8jld\" (UID: \"ecdd722c-78ea-4e3e-bd54-31e53b1177da\") " pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.106076 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.518598 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8jld"] Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.755749 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgpdn"] Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.756721 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.761906 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.765144 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgpdn"] Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.802740 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk7sw\" (UniqueName: \"kubernetes.io/projected/f35183da-e8fc-4399-a3fa-20b466813639-kube-api-access-mk7sw\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.802823 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-utilities\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.802844 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-catalog-content\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.905163 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk7sw\" (UniqueName: \"kubernetes.io/projected/f35183da-e8fc-4399-a3fa-20b466813639-kube-api-access-mk7sw\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.905288 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-utilities\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.905319 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-catalog-content\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.906045 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-catalog-content\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.906315 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f35183da-e8fc-4399-a3fa-20b466813639-utilities\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:01 crc kubenswrapper[4800]: I0122 00:11:01.925229 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk7sw\" (UniqueName: \"kubernetes.io/projected/f35183da-e8fc-4399-a3fa-20b466813639-kube-api-access-mk7sw\") pod \"certified-operators-rgpdn\" (UID: \"f35183da-e8fc-4399-a3fa-20b466813639\") " pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:02 crc kubenswrapper[4800]: I0122 00:11:02.111698 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:02 crc kubenswrapper[4800]: I0122 00:11:02.191034 4800 generic.go:334] "Generic (PLEG): container finished" podID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" containerID="68fe47fc9cb7d218416cef07a842e3894b43d900296a28ca5d06438b74d78c14" exitCode=0 Jan 22 00:11:02 crc kubenswrapper[4800]: I0122 00:11:02.191214 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8jld" event={"ID":"ecdd722c-78ea-4e3e-bd54-31e53b1177da","Type":"ContainerDied","Data":"68fe47fc9cb7d218416cef07a842e3894b43d900296a28ca5d06438b74d78c14"} Jan 22 00:11:02 crc kubenswrapper[4800]: I0122 00:11:02.191381 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8jld" event={"ID":"ecdd722c-78ea-4e3e-bd54-31e53b1177da","Type":"ContainerStarted","Data":"00a206f01a852e530bdd750e346f8dbd92e0c0ce8de53cfb6e47020b303b8d57"} Jan 22 00:11:02 crc kubenswrapper[4800]: I0122 00:11:02.306218 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgpdn"] Jan 22 00:11:02 crc kubenswrapper[4800]: W0122 00:11:02.324035 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf35183da_e8fc_4399_a3fa_20b466813639.slice/crio-d1e2433be937036eea6b36b5e6de1f14bac2e377cbfe91b90c08b7b8ad7ff811 WatchSource:0}: Error finding container d1e2433be937036eea6b36b5e6de1f14bac2e377cbfe91b90c08b7b8ad7ff811: Status 404 returned error can't find the container with id d1e2433be937036eea6b36b5e6de1f14bac2e377cbfe91b90c08b7b8ad7ff811 Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.150967 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.152880 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.155341 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.161228 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.196525 4800 generic.go:334] "Generic (PLEG): container finished" podID="f35183da-e8fc-4399-a3fa-20b466813639" containerID="82651b02971fd342429bc26b115ae370744de53fcb46b8f09bc74ac1920bc4ad" exitCode=0 Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.196573 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgpdn" event={"ID":"f35183da-e8fc-4399-a3fa-20b466813639","Type":"ContainerDied","Data":"82651b02971fd342429bc26b115ae370744de53fcb46b8f09bc74ac1920bc4ad"} Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.196618 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgpdn" event={"ID":"f35183da-e8fc-4399-a3fa-20b466813639","Type":"ContainerStarted","Data":"d1e2433be937036eea6b36b5e6de1f14bac2e377cbfe91b90c08b7b8ad7ff811"} Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.198528 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8jld" event={"ID":"ecdd722c-78ea-4e3e-bd54-31e53b1177da","Type":"ContainerStarted","Data":"7d855087b0625aba5694454810b27d8f6445700db796cc2c636fcbbff5eda485"} Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.224087 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.224142 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.224194 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzjll\" (UniqueName: \"kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.325139 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.325213 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.325275 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzjll\" (UniqueName: \"kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.326104 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.326721 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.346089 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzjll\" (UniqueName: \"kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll\") pod \"redhat-marketplace-kckpd\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.471911 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:03 crc kubenswrapper[4800]: I0122 00:11:03.661446 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:11:03 crc kubenswrapper[4800]: W0122 00:11:03.701496 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f77d02c_3957_4ebe_bdc5_e17d73bf4eb7.slice/crio-8d31728e6b52f4706e06ec8ffcafa356c10affe6e5921c1e8eb1f2e3bb7c7035 WatchSource:0}: Error finding container 8d31728e6b52f4706e06ec8ffcafa356c10affe6e5921c1e8eb1f2e3bb7c7035: Status 404 returned error can't find the container with id 8d31728e6b52f4706e06ec8ffcafa356c10affe6e5921c1e8eb1f2e3bb7c7035 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.151832 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h7qwp"] Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.153564 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.156194 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.160364 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7qwp"] Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.204222 4800 generic.go:334] "Generic (PLEG): container finished" podID="f35183da-e8fc-4399-a3fa-20b466813639" containerID="f23080d39d0627028e2fb5a01d6a2aa737e8afaecb19863942b00598e62062fe" exitCode=0 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.204798 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgpdn" event={"ID":"f35183da-e8fc-4399-a3fa-20b466813639","Type":"ContainerDied","Data":"f23080d39d0627028e2fb5a01d6a2aa737e8afaecb19863942b00598e62062fe"} Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.207691 4800 generic.go:334] "Generic (PLEG): container finished" podID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" containerID="7d855087b0625aba5694454810b27d8f6445700db796cc2c636fcbbff5eda485" exitCode=0 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.207753 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8jld" event={"ID":"ecdd722c-78ea-4e3e-bd54-31e53b1177da","Type":"ContainerDied","Data":"7d855087b0625aba5694454810b27d8f6445700db796cc2c636fcbbff5eda485"} Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.210358 4800 generic.go:334] "Generic (PLEG): container finished" podID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerID="f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081" exitCode=0 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.210402 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerDied","Data":"f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081"} Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.210577 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerStarted","Data":"8d31728e6b52f4706e06ec8ffcafa356c10affe6e5921c1e8eb1f2e3bb7c7035"} Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.236963 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-utilities\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.237273 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.237307 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-catalog-content\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.302612 4800 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303324 4800 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303497 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303577 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444" gracePeriod=15 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303601 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb" gracePeriod=15 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303619 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd" gracePeriod=15 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303642 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d" gracePeriod=15 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.303623 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174" gracePeriod=15 Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.305900 4800 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306040 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306057 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306066 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306073 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306080 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306087 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306100 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306106 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306116 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306122 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306131 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306136 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.306147 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306153 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306239 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306252 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306259 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306267 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306275 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.306418 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.338541 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.338678 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-catalog-content\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.338804 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-utilities\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.339377 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-utilities\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.339741 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/994f0255-587e-4ad7-a84f-c4c8a0e06184-catalog-content\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440035 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440103 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440121 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440140 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440153 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440182 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440322 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.440433 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541619 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541677 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541717 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541748 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541757 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541787 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541824 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541839 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541821 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541894 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541938 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.541968 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.542044 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.542066 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.542137 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.542141 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:04 crc kubenswrapper[4800]: E0122 00:11:04.597242 4800 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.242:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-x8jld.188ce51c967f4ade openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-x8jld,UID:ecdd722c-78ea-4e3e-bd54-31e53b1177da,APIVersion:v1,ResourceVersion:29564,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 387ms (387ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 00:11:04.596531934 +0000 UTC m=+240.076834972,LastTimestamp:2026-01-22 00:11:04.596531934 +0000 UTC m=+240.076834972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 00:11:04 crc kubenswrapper[4800]: I0122 00:11:04.820574 4800 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.217425 4800 generic.go:334] "Generic (PLEG): container finished" podID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerID="6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed" exitCode=0 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.217477 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerDied","Data":"6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed"} Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.218284 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.221502 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.223345 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.224292 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174" exitCode=0 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.224324 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb" exitCode=0 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.224333 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd" exitCode=0 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.224343 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d" exitCode=2 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.224418 4800 scope.go:117] "RemoveContainer" containerID="79a23f2501a0f838f5d79a900881cd6d063c656bd4b328d1061db1db8735a003" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.228144 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgpdn" event={"ID":"f35183da-e8fc-4399-a3fa-20b466813639","Type":"ContainerStarted","Data":"530d5f99d2d5e9bfdf697408e134535b2f7918776de0fde5f40377fcd58a8030"} Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.228713 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.230016 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.232425 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8jld" event={"ID":"ecdd722c-78ea-4e3e-bd54-31e53b1177da","Type":"ContainerStarted","Data":"8acdfadac2c3c7114915fabecc8e64423e0e225de794737ebe089b5492bd832c"} Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.233954 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.234556 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.235158 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.235777 4800 generic.go:334] "Generic (PLEG): container finished" podID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" containerID="30fbca53cfb0e130f774103597b5e5b3452d5487268655fd04b13a7546bb70f5" exitCode=0 Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.235831 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5dfdfa65-9944-4064-a852-b6bc9833ffa8","Type":"ContainerDied","Data":"30fbca53cfb0e130f774103597b5e5b3452d5487268655fd04b13a7546bb70f5"} Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.236490 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.236768 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.237078 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:05 crc kubenswrapper[4800]: I0122 00:11:05.237369 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.243532 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.248029 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerStarted","Data":"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de"} Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.248770 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.249012 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.249312 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.249770 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.675913 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.676842 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.677269 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.677655 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.678251 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.681798 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.682726 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.683297 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.683708 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.684063 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.684313 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.684652 4800 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771056 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771116 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir\") pod \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771134 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock\") pod \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771148 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771178 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771254 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access\") pod \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\" (UID: \"5dfdfa65-9944-4064-a852-b6bc9833ffa8\") " Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771252 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771293 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771252 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock" (OuterVolumeSpecName: "var-lock") pod "5dfdfa65-9944-4064-a852-b6bc9833ffa8" (UID: "5dfdfa65-9944-4064-a852-b6bc9833ffa8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771253 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5dfdfa65-9944-4064-a852-b6bc9833ffa8" (UID: "5dfdfa65-9944-4064-a852-b6bc9833ffa8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771352 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771472 4800 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771484 4800 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771493 4800 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771503 4800 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.771511 4800 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.778826 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5dfdfa65-9944-4064-a852-b6bc9833ffa8" (UID: "5dfdfa65-9944-4064-a852-b6bc9833ffa8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.824007 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 22 00:11:06 crc kubenswrapper[4800]: I0122 00:11:06.872630 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dfdfa65-9944-4064-a852-b6bc9833ffa8-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.256581 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.257263 4800 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444" exitCode=0 Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.257352 4800 scope.go:117] "RemoveContainer" containerID="5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.257350 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.257956 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.258180 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.258519 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.258726 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.259445 4800 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.259859 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.259865 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5dfdfa65-9944-4064-a852-b6bc9833ffa8","Type":"ContainerDied","Data":"b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b"} Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.259918 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4ac4abd4abc7cbb10b14b6edbd36dd0e890c23ce0414b49c0b069d6114c635b" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.262012 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.262263 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.262520 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.262790 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.263032 4800 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.265523 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.265934 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.266167 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.266390 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.266601 4800 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.279868 4800 scope.go:117] "RemoveContainer" containerID="cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.299280 4800 scope.go:117] "RemoveContainer" containerID="1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.317160 4800 scope.go:117] "RemoveContainer" containerID="f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.333472 4800 scope.go:117] "RemoveContainer" containerID="88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.351223 4800 scope.go:117] "RemoveContainer" containerID="b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.371589 4800 scope.go:117] "RemoveContainer" containerID="5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.372605 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\": container with ID starting with 5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174 not found: ID does not exist" containerID="5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.372649 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174"} err="failed to get container status \"5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\": rpc error: code = NotFound desc = could not find container \"5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174\": container with ID starting with 5f39fdf848c4ef50152560eb6fe4261350d6d7f3553016c9d75f0407a822c174 not found: ID does not exist" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.372679 4800 scope.go:117] "RemoveContainer" containerID="cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.374012 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\": container with ID starting with cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb not found: ID does not exist" containerID="cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.374069 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb"} err="failed to get container status \"cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\": rpc error: code = NotFound desc = could not find container \"cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb\": container with ID starting with cc279d613bf6148fc19b6fb62787054bd53a7ad3f1c0c8be9653ea9778cc7adb not found: ID does not exist" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.374113 4800 scope.go:117] "RemoveContainer" containerID="1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.376486 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\": container with ID starting with 1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd not found: ID does not exist" containerID="1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.376531 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd"} err="failed to get container status \"1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\": rpc error: code = NotFound desc = could not find container \"1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd\": container with ID starting with 1d6d227f1137209f8476d8fd211884c370e13a152988dec537fbcace132afebd not found: ID does not exist" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.376565 4800 scope.go:117] "RemoveContainer" containerID="f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.377314 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\": container with ID starting with f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d not found: ID does not exist" containerID="f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.377349 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d"} err="failed to get container status \"f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\": rpc error: code = NotFound desc = could not find container \"f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d\": container with ID starting with f87e2b1dd4faf1e505bcdd2999e6a3a92a8bdd0ce91368f63f2c98d36b3cc24d not found: ID does not exist" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.377370 4800 scope.go:117] "RemoveContainer" containerID="88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.378222 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\": container with ID starting with 88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444 not found: ID does not exist" containerID="88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.378251 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444"} err="failed to get container status \"88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\": rpc error: code = NotFound desc = could not find container \"88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444\": container with ID starting with 88ff35137b47e5c25c1ff5034f0a85f77aa9a5a2e31108745fbc82f8acf7a444 not found: ID does not exist" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.378269 4800 scope.go:117] "RemoveContainer" containerID="b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27" Jan 22 00:11:07 crc kubenswrapper[4800]: E0122 00:11:07.378931 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\": container with ID starting with b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27 not found: ID does not exist" containerID="b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27" Jan 22 00:11:07 crc kubenswrapper[4800]: I0122 00:11:07.378965 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27"} err="failed to get container status \"b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\": rpc error: code = NotFound desc = could not find container \"b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27\": container with ID starting with b45b18c748b5fdf536370beeca07096ac8ad3f06bbcfdb57bff4a86eeddfcc27 not found: ID does not exist" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.340718 4800 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:09 crc kubenswrapper[4800]: I0122 00:11:09.341903 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.343969 4800 projected.go:194] Error preparing data for projected volume kube-api-access-mhfzq for pod openshift-marketplace/community-operators-h7qwp: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.344058 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq podName:994f0255-587e-4ad7-a84f-c4c8a0e06184 nodeName:}" failed. No retries permitted until 2026-01-22 00:11:09.844032056 +0000 UTC m=+245.324335084 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mhfzq" (UniqueName: "kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq") pod "community-operators-h7qwp" (UID: "994f0255-587e-4ad7-a84f-c4c8a0e06184") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.515110 4800 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.515963 4800 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.516398 4800 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.516735 4800 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.517100 4800 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:09 crc kubenswrapper[4800]: I0122 00:11:09.517137 4800 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.517433 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="200ms" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.718623 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="400ms" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.879452 4800 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" volumeName="registry-storage" Jan 22 00:11:09 crc kubenswrapper[4800]: I0122 00:11:09.912832 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.914143 4800 projected.go:194] Error preparing data for projected volume kube-api-access-mhfzq for pod openshift-marketplace/community-operators-h7qwp: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:09 crc kubenswrapper[4800]: E0122 00:11:09.914228 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq podName:994f0255-587e-4ad7-a84f-c4c8a0e06184 nodeName:}" failed. No retries permitted until 2026-01-22 00:11:10.914205162 +0000 UTC m=+246.394508200 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-mhfzq" (UniqueName: "kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq") pod "community-operators-h7qwp" (UID: "994f0255-587e-4ad7-a84f-c4c8a0e06184") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:10 crc kubenswrapper[4800]: E0122 00:11:10.120096 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="800ms" Jan 22 00:11:10 crc kubenswrapper[4800]: I0122 00:11:10.284580 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"950ced80a2e73c792f4979e20f0b20d10f9db5f9140160d3ff7461019bf46a3f"} Jan 22 00:11:10 crc kubenswrapper[4800]: E0122 00:11:10.921622 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="1.6s" Jan 22 00:11:10 crc kubenswrapper[4800]: I0122 00:11:10.924541 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:10 crc kubenswrapper[4800]: E0122 00:11:10.925056 4800 projected.go:194] Error preparing data for projected volume kube-api-access-mhfzq for pod openshift-marketplace/community-operators-h7qwp: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:10 crc kubenswrapper[4800]: E0122 00:11:10.925127 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq podName:994f0255-587e-4ad7-a84f-c4c8a0e06184 nodeName:}" failed. No retries permitted until 2026-01-22 00:11:12.925106941 +0000 UTC m=+248.405409979 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-mhfzq" (UniqueName: "kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq") pod "community-operators-h7qwp" (UID: "994f0255-587e-4ad7-a84f-c4c8a0e06184") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.106862 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.107260 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:11 crc kubenswrapper[4800]: E0122 00:11:11.150055 4800 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.242:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-x8jld.188ce51c967f4ade openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-x8jld,UID:ecdd722c-78ea-4e3e-bd54-31e53b1177da,APIVersion:v1,ResourceVersion:29564,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 387ms (387ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 00:11:04.596531934 +0000 UTC m=+240.076834972,LastTimestamp:2026-01-22 00:11:04.596531934 +0000 UTC m=+240.076834972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.157746 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.158182 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.158341 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.158510 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.158653 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.299300 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9"} Jan 22 00:11:11 crc kubenswrapper[4800]: E0122 00:11:11.299967 4800 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.300192 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.300452 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.300628 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.300840 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.334934 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x8jld" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.335558 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.336038 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.336337 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:11 crc kubenswrapper[4800]: I0122 00:11:11.336621 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.113011 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.113070 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.155913 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.157140 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.157539 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.157721 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.158029 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: E0122 00:11:12.304041 4800 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.341418 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgpdn" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.341974 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.342297 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.342734 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.342949 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:12 crc kubenswrapper[4800]: E0122 00:11:12.523425 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="3.2s" Jan 22 00:11:12 crc kubenswrapper[4800]: I0122 00:11:12.948111 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:12 crc kubenswrapper[4800]: E0122 00:11:12.948540 4800 projected.go:194] Error preparing data for projected volume kube-api-access-mhfzq for pod openshift-marketplace/community-operators-h7qwp: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:12 crc kubenswrapper[4800]: E0122 00:11:12.948593 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq podName:994f0255-587e-4ad7-a84f-c4c8a0e06184 nodeName:}" failed. No retries permitted until 2026-01-22 00:11:16.948577805 +0000 UTC m=+252.428880843 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-mhfzq" (UniqueName: "kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq") pod "community-operators-h7qwp" (UID: "994f0255-587e-4ad7-a84f-c4c8a0e06184") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.473134 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.473185 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.519177 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.519860 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.520384 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.520845 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:13 crc kubenswrapper[4800]: I0122 00:11:13.521098 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.355593 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.356235 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.356612 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.357012 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.357292 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.822340 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.823054 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.823298 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:14 crc kubenswrapper[4800]: I0122 00:11:14.823557 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:15 crc kubenswrapper[4800]: E0122 00:11:15.724041 4800 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.242:6443: connect: connection refused" interval="6.4s" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.817940 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.818681 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.819011 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.819270 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.819495 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.833020 4800 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.833052 4800 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:15 crc kubenswrapper[4800]: E0122 00:11:15.833404 4800 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:15 crc kubenswrapper[4800]: I0122 00:11:15.833955 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:15 crc kubenswrapper[4800]: W0122 00:11:15.858572 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-52f5e1e821cad95948a5df1bbbbb025e1817b766a7011106809ecdfadb9b36cc WatchSource:0}: Error finding container 52f5e1e821cad95948a5df1bbbbb025e1817b766a7011106809ecdfadb9b36cc: Status 404 returned error can't find the container with id 52f5e1e821cad95948a5df1bbbbb025e1817b766a7011106809ecdfadb9b36cc Jan 22 00:11:16 crc kubenswrapper[4800]: I0122 00:11:16.351620 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52f5e1e821cad95948a5df1bbbbb025e1817b766a7011106809ecdfadb9b36cc"} Jan 22 00:11:16 crc kubenswrapper[4800]: I0122 00:11:16.953285 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:16 crc kubenswrapper[4800]: E0122 00:11:16.954195 4800 projected.go:194] Error preparing data for projected volume kube-api-access-mhfzq for pod openshift-marketplace/community-operators-h7qwp: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:16 crc kubenswrapper[4800]: E0122 00:11:16.954299 4800 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq podName:994f0255-587e-4ad7-a84f-c4c8a0e06184 nodeName:}" failed. No retries permitted until 2026-01-22 00:11:24.954273318 +0000 UTC m=+260.434576356 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-mhfzq" (UniqueName: "kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq") pod "community-operators-h7qwp" (UID: "994f0255-587e-4ad7-a84f-c4c8a0e06184") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/community-operators/token": dial tcp 38.129.56.242:6443: connect: connection refused Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.038631 4800 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.038711 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.366201 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.366672 4800 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d" exitCode=1 Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.366770 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d"} Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.367394 4800 scope.go:117] "RemoveContainer" containerID="8a300ad1345c67b868f11d6a46b27c3f050f4f8c7b52aa3d542d6982d574a94d" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.368084 4800 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.368595 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369174 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369396 4800 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="4a03b3192892069f743847bac464924bca983733180fd726dd1d91a6ee616d5c" exitCode=0 Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369433 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"4a03b3192892069f743847bac464924bca983733180fd726dd1d91a6ee616d5c"} Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369749 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369754 4800 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.369921 4800 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.370285 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: E0122 00:11:18.370579 4800 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.370991 4800 status_manager.go:851] "Failed to get status for pod" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.371451 4800 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.371868 4800 status_manager.go:851] "Failed to get status for pod" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" pod="openshift-marketplace/redhat-marketplace-kckpd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-kckpd\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.372298 4800 status_manager.go:851] "Failed to get status for pod" podUID="ecdd722c-78ea-4e3e-bd54-31e53b1177da" pod="openshift-marketplace/redhat-operators-x8jld" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-x8jld\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:18 crc kubenswrapper[4800]: I0122 00:11:18.372613 4800 status_manager.go:851] "Failed to get status for pod" podUID="f35183da-e8fc-4399-a3fa-20b466813639" pod="openshift-marketplace/certified-operators-rgpdn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rgpdn\": dial tcp 38.129.56.242:6443: connect: connection refused" Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.386598 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.387331 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"91edffc62063839f67c33cdd0d61fdf5987b1360a1c582e3c8b7201c25aeedc5"} Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.390151 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d5fb7c0a173e83d4c0b511092a2b3ae4d69d9c11768c9bf3a0bb53e9caf6a5dc"} Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.390198 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"067efa1b5408b6aa9e1c8402de546e5a53b8ca477034a6c4f1b374c8dce554d4"} Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.390211 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"85d396460d4b1801c9e86a4bcf2e5392720dbd7368a39cf71f00306b9ad738db"} Jan 22 00:11:19 crc kubenswrapper[4800]: I0122 00:11:19.833473 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.397839 4800 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.397870 4800 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.398079 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9f50639c3ec652f24f212b7ea122186de6119068acf1b9bef5bef83fed55bb1e"} Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.398115 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fe7272dc0d5578f99257e67c336367af8b4c0ee889ca2df4a951eda02b1c22e1"} Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.398135 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.835414 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.835456 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.841042 4800 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]log ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]etcd ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/generic-apiserver-start-informers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-filter ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-apiextensions-informers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-apiextensions-controllers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/crd-informer-synced ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-system-namespaces-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 22 00:11:20 crc kubenswrapper[4800]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/bootstrap-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/start-kube-aggregator-informers ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-registration-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-discovery-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]autoregister-completion ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-openapi-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 22 00:11:20 crc kubenswrapper[4800]: livez check failed Jan 22 00:11:20 crc kubenswrapper[4800]: I0122 00:11:20.841110 4800 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 00:11:21 crc kubenswrapper[4800]: I0122 00:11:21.937492 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" podUID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" containerName="oauth-openshift" containerID="cri-o://c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3" gracePeriod=15 Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.288514 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.409553 4800 generic.go:334] "Generic (PLEG): container finished" podID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" containerID="c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3" exitCode=0 Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.409612 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" event={"ID":"666d239f-66e7-48de-bc1c-6d0c1ab4f774","Type":"ContainerDied","Data":"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3"} Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.409616 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.409665 4800 scope.go:117] "RemoveContainer" containerID="c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.409651 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bq4b8" event={"ID":"666d239f-66e7-48de-bc1c-6d0c1ab4f774","Type":"ContainerDied","Data":"793cca827c9b2873f7ad4aa048dad5f4ea9ba15d37da91a9ad6d65cb263ac4d3"} Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420631 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420719 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420755 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420786 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420804 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420827 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420868 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420906 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420929 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420953 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.420979 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.421007 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.421030 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2c48\" (UniqueName: \"kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.421054 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session\") pod \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\" (UID: \"666d239f-66e7-48de-bc1c-6d0c1ab4f774\") " Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.421861 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.421911 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.422006 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.422043 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.422089 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.427324 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.427972 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.428192 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.428185 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48" (OuterVolumeSpecName: "kube-api-access-k2c48") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "kube-api-access-k2c48". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.429284 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.429773 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.430390 4800 scope.go:117] "RemoveContainer" containerID="c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3" Jan 22 00:11:22 crc kubenswrapper[4800]: E0122 00:11:22.431131 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3\": container with ID starting with c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3 not found: ID does not exist" containerID="c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.431162 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3"} err="failed to get container status \"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3\": rpc error: code = NotFound desc = could not find container \"c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3\": container with ID starting with c6a90ca5aaf5cff67405dceec24282362cd7b8dac348fab811e07a4e9d322eb3 not found: ID does not exist" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.431926 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.440079 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.440360 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "666d239f-66e7-48de-bc1c-6d0c1ab4f774" (UID: "666d239f-66e7-48de-bc1c-6d0c1ab4f774"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522516 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2c48\" (UniqueName: \"kubernetes.io/projected/666d239f-66e7-48de-bc1c-6d0c1ab4f774-kube-api-access-k2c48\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522557 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522567 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522579 4800 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522590 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522600 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522609 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522618 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522627 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522637 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522645 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522656 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522663 4800 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/666d239f-66e7-48de-bc1c-6d0c1ab4f774-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:22 crc kubenswrapper[4800]: I0122 00:11:22.522671 4800 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/666d239f-66e7-48de-bc1c-6d0c1ab4f774-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:24 crc kubenswrapper[4800]: I0122 00:11:24.962851 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:24 crc kubenswrapper[4800]: I0122 00:11:24.992563 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhfzq\" (UniqueName: \"kubernetes.io/projected/994f0255-587e-4ad7-a84f-c4c8a0e06184-kube-api-access-mhfzq\") pod \"community-operators-h7qwp\" (UID: \"994f0255-587e-4ad7-a84f-c4c8a0e06184\") " pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.179485 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.188610 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.410314 4800 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.428183 4800 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:11:18Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T00:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85d396460d4b1801c9e86a4bcf2e5392720dbd7368a39cf71f00306b9ad738db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5fb7c0a173e83d4c0b511092a2b3ae4d69d9c11768c9bf3a0bb53e9caf6a5dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://067efa1b5408b6aa9e1c8402de546e5a53b8ca477034a6c4f1b374c8dce554d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f50639c3ec652f24f212b7ea122186de6119068acf1b9bef5bef83fed55bb1e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7272dc0d5578f99257e67c336367af8b4c0ee889ca2df4a951eda02b1c22e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T00:11:19Z\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a03b3192892069f743847bac464924bca983733180fd726dd1d91a6ee616d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a03b3192892069f743847bac464924bca983733180fd726dd1d91a6ee616d5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T00:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T00:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}]}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Pod \"kube-apiserver-crc\" is invalid: metadata.uid: Invalid value: \"80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2\": field is immutable" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.436563 4800 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.436589 4800 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="80c54dc4-bdbc-4c60-a6fa-feea7fe5d9d2" Jan 22 00:11:25 crc kubenswrapper[4800]: I0122 00:11:25.477001 4800 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4abaf1e2-9c0f-4b01-a8cc-0fe422592037" Jan 22 00:11:25 crc kubenswrapper[4800]: W0122 00:11:25.675693 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod994f0255_587e_4ad7_a84f_c4c8a0e06184.slice/crio-79e9c0c7ae8204c7a906cb9588c9ce4c0f52c52f0f4fa50c156711d5f6063cef WatchSource:0}: Error finding container 79e9c0c7ae8204c7a906cb9588c9ce4c0f52c52f0f4fa50c156711d5f6063cef: Status 404 returned error can't find the container with id 79e9c0c7ae8204c7a906cb9588c9ce4c0f52c52f0f4fa50c156711d5f6063cef Jan 22 00:11:25 crc kubenswrapper[4800]: E0122 00:11:25.692816 4800 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Jan 22 00:11:26 crc kubenswrapper[4800]: I0122 00:11:26.465348 4800 generic.go:334] "Generic (PLEG): container finished" podID="994f0255-587e-4ad7-a84f-c4c8a0e06184" containerID="852d0ae86af578907385f876f5c8de6f42b7d2bb0f15ad70dfcdb35675e308f7" exitCode=0 Jan 22 00:11:26 crc kubenswrapper[4800]: I0122 00:11:26.465394 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7qwp" event={"ID":"994f0255-587e-4ad7-a84f-c4c8a0e06184","Type":"ContainerDied","Data":"852d0ae86af578907385f876f5c8de6f42b7d2bb0f15ad70dfcdb35675e308f7"} Jan 22 00:11:26 crc kubenswrapper[4800]: I0122 00:11:26.465420 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7qwp" event={"ID":"994f0255-587e-4ad7-a84f-c4c8a0e06184","Type":"ContainerStarted","Data":"79e9c0c7ae8204c7a906cb9588c9ce4c0f52c52f0f4fa50c156711d5f6063cef"} Jan 22 00:11:26 crc kubenswrapper[4800]: I0122 00:11:26.980637 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:11:26 crc kubenswrapper[4800]: I0122 00:11:26.984148 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:11:27 crc kubenswrapper[4800]: I0122 00:11:27.471590 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7qwp" event={"ID":"994f0255-587e-4ad7-a84f-c4c8a0e06184","Type":"ContainerStarted","Data":"88d8e29bd1144119e8ebc82d28e5cf0ecfa557d7aa7fded0b0e1cb0a16b699b3"} Jan 22 00:11:28 crc kubenswrapper[4800]: I0122 00:11:28.479929 4800 generic.go:334] "Generic (PLEG): container finished" podID="994f0255-587e-4ad7-a84f-c4c8a0e06184" containerID="88d8e29bd1144119e8ebc82d28e5cf0ecfa557d7aa7fded0b0e1cb0a16b699b3" exitCode=0 Jan 22 00:11:28 crc kubenswrapper[4800]: I0122 00:11:28.480335 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7qwp" event={"ID":"994f0255-587e-4ad7-a84f-c4c8a0e06184","Type":"ContainerDied","Data":"88d8e29bd1144119e8ebc82d28e5cf0ecfa557d7aa7fded0b0e1cb0a16b699b3"} Jan 22 00:11:29 crc kubenswrapper[4800]: I0122 00:11:29.488832 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7qwp" event={"ID":"994f0255-587e-4ad7-a84f-c4c8a0e06184","Type":"ContainerStarted","Data":"90b9d4174ba9697fa8c6d22e3c5ef3a5d1ca1b0b0ed9ddf59bc60d659a4fa46e"} Jan 22 00:11:29 crc kubenswrapper[4800]: I0122 00:11:29.839603 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 00:11:35 crc kubenswrapper[4800]: I0122 00:11:35.190234 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:35 crc kubenswrapper[4800]: I0122 00:11:35.190646 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:35 crc kubenswrapper[4800]: I0122 00:11:35.237082 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:35 crc kubenswrapper[4800]: I0122 00:11:35.562297 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h7qwp" Jan 22 00:11:35 crc kubenswrapper[4800]: I0122 00:11:35.922349 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 00:11:36 crc kubenswrapper[4800]: I0122 00:11:36.229408 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 00:11:36 crc kubenswrapper[4800]: I0122 00:11:36.547160 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 00:11:36 crc kubenswrapper[4800]: I0122 00:11:36.604627 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 00:11:36 crc kubenswrapper[4800]: I0122 00:11:36.797977 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 00:11:36 crc kubenswrapper[4800]: I0122 00:11:36.802456 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.069132 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.079439 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.152847 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.196591 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.236976 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.349018 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.608169 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.622562 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.732845 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.813461 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 00:11:37 crc kubenswrapper[4800]: I0122 00:11:37.920429 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.182278 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.188189 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.225506 4800 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.227114 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h7qwp" podStartSLOduration=31.833127666 podStartE2EDuration="34.227098693s" podCreationTimestamp="2026-01-22 00:11:04 +0000 UTC" firstStartedPulling="2026-01-22 00:11:26.466763621 +0000 UTC m=+261.947066659" lastFinishedPulling="2026-01-22 00:11:28.860734648 +0000 UTC m=+264.341037686" observedRunningTime="2026-01-22 00:11:29.508239158 +0000 UTC m=+264.988542196" watchObservedRunningTime="2026-01-22 00:11:38.227098693 +0000 UTC m=+273.707401721" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.227554 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgpdn" podStartSLOduration=35.723023987 podStartE2EDuration="37.227547855s" podCreationTimestamp="2026-01-22 00:11:01 +0000 UTC" firstStartedPulling="2026-01-22 00:11:03.197791306 +0000 UTC m=+238.678094344" lastFinishedPulling="2026-01-22 00:11:04.702315174 +0000 UTC m=+240.182618212" observedRunningTime="2026-01-22 00:11:25.338018909 +0000 UTC m=+260.818321947" watchObservedRunningTime="2026-01-22 00:11:38.227547855 +0000 UTC m=+273.707850893" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.228198 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kckpd" podStartSLOduration=33.8124414 podStartE2EDuration="35.228193162s" podCreationTimestamp="2026-01-22 00:11:03 +0000 UTC" firstStartedPulling="2026-01-22 00:11:04.211615264 +0000 UTC m=+239.691918302" lastFinishedPulling="2026-01-22 00:11:05.627367036 +0000 UTC m=+241.107670064" observedRunningTime="2026-01-22 00:11:25.255958652 +0000 UTC m=+260.736261700" watchObservedRunningTime="2026-01-22 00:11:38.228193162 +0000 UTC m=+273.708496190" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.228731 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x8jld" podStartSLOduration=35.835661134 podStartE2EDuration="38.228723677s" podCreationTimestamp="2026-01-22 00:11:00 +0000 UTC" firstStartedPulling="2026-01-22 00:11:02.20345485 +0000 UTC m=+237.683757888" lastFinishedPulling="2026-01-22 00:11:04.596517393 +0000 UTC m=+240.076820431" observedRunningTime="2026-01-22 00:11:25.287025723 +0000 UTC m=+260.767328761" watchObservedRunningTime="2026-01-22 00:11:38.228723677 +0000 UTC m=+273.709026715" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.229787 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bq4b8","openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.229831 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.229847 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7qwp"] Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.238793 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.242649 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.253023 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.253005159 podStartE2EDuration="13.253005159s" podCreationTimestamp="2026-01-22 00:11:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:11:38.249644518 +0000 UTC m=+273.729947556" watchObservedRunningTime="2026-01-22 00:11:38.253005159 +0000 UTC m=+273.733308197" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.321177 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.375633 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.420534 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.526128 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.699837 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.776607 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.820286 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.824798 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" path="/var/lib/kubelet/pods/666d239f-66e7-48de-bc1c-6d0c1ab4f774/volumes" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.856044 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 00:11:38 crc kubenswrapper[4800]: I0122 00:11:38.928591 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.064477 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.118780 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.131602 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.195947 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.255073 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.312935 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.331114 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.365951 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.382289 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.391540 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.426729 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.501456 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.511634 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.554804 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.573049 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.579267 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-55896b6b9d-phdx7"] Jan 22 00:11:39 crc kubenswrapper[4800]: E0122 00:11:39.579473 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" containerName="oauth-openshift" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.579486 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" containerName="oauth-openshift" Jan 22 00:11:39 crc kubenswrapper[4800]: E0122 00:11:39.579498 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" containerName="installer" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.579503 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" containerName="installer" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.579600 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="666d239f-66e7-48de-bc1c-6d0c1ab4f774" containerName="oauth-openshift" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.579613 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dfdfa65-9944-4064-a852-b6bc9833ffa8" containerName="installer" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.580097 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.582558 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.582814 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.582969 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.583382 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.583703 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.583737 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.583832 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.583702 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.584068 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.584090 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.584300 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.584321 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.592629 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.594248 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.606963 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.625553 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.733446 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747365 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-error\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747421 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-login\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747448 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-dir\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747471 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-router-certs\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747489 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-policies\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747510 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-service-ca\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747536 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747555 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747573 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747599 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-session\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747614 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmlbk\" (UniqueName: \"kubernetes.io/projected/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-kube-api-access-qmlbk\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747630 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747646 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.747661 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.766637 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.775530 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.795644 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848278 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848344 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-session\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848361 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmlbk\" (UniqueName: \"kubernetes.io/projected/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-kube-api-access-qmlbk\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848380 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848398 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848430 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848473 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-error\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848494 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-login\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848510 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-dir\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848529 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-policies\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848545 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-router-certs\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848560 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-service-ca\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848585 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.848601 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.849375 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.849425 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-dir\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.849835 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-audit-policies\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.850853 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-service-ca\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.851138 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.854732 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-login\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.855056 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.855986 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-template-error\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.856265 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-router-certs\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.856496 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.857250 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.863049 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.864471 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-v4-0-config-system-session\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.877845 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmlbk\" (UniqueName: \"kubernetes.io/projected/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01-kube-api-access-qmlbk\") pod \"oauth-openshift-55896b6b9d-phdx7\" (UID: \"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\") " pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:39 crc kubenswrapper[4800]: I0122 00:11:39.900325 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.062119 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.131559 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.241140 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.265546 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.276347 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.367082 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.425947 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.448175 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.458711 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.545713 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.579947 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.733875 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.761742 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.839020 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.843099 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.845464 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.848092 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.886508 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 00:11:40 crc kubenswrapper[4800]: I0122 00:11:40.988816 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.108546 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.203299 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.533152 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.578644 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.581412 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.644648 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.703230 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.819592 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.910101 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.918273 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.947820 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 00:11:41 crc kubenswrapper[4800]: I0122 00:11:41.989795 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.105603 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.143735 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.233338 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.255387 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.529614 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.553759 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.603912 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.604567 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.639453 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.740183 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.764423 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.776549 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.810384 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.813543 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.837493 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.843525 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.916863 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 00:11:42 crc kubenswrapper[4800]: I0122 00:11:42.941138 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.067534 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.072310 4800 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.373988 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.403657 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.416424 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.425915 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.481619 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.517120 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.539362 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.546515 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.679252 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.748038 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.808260 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.854828 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.868840 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.925550 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 00:11:43 crc kubenswrapper[4800]: I0122 00:11:43.968720 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.041468 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.128146 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.196635 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.275612 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.281483 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.288736 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.293099 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.306742 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.371431 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.378626 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.402383 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.405926 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.523617 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.542944 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.555030 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.640577 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.667565 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.673609 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.725386 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.773085 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.798752 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.817361 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.916681 4800 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.924668 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 00:11:44 crc kubenswrapper[4800]: I0122 00:11:44.987001 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.030981 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.250766 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.351729 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.547182 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.585817 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.594175 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.674380 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.684178 4800 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.798558 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.840703 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.943615 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 00:11:45 crc kubenswrapper[4800]: I0122 00:11:45.963365 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.001451 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.006094 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.015657 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.025875 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.152781 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.168012 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.293812 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.294394 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.318402 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.339475 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.372432 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.396207 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.412222 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.499020 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.525043 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.556976 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.645690 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.708220 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.790775 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.821965 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.913977 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.965324 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 00:11:46 crc kubenswrapper[4800]: I0122 00:11:46.971950 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.009828 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.012796 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.131630 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.162550 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.165664 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.283748 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.374728 4800 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.433324 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.453238 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.522797 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.632751 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.640731 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.758559 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.915854 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.950337 4800 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.951525 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9" gracePeriod=5 Jan 22 00:11:47 crc kubenswrapper[4800]: I0122 00:11:47.959492 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.052546 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.096438 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.137062 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.173703 4800 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.186659 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.208395 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.227755 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.254589 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.357101 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.440342 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.444402 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.563119 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.607372 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.634055 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.787486 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.798563 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.879051 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 00:11:48 crc kubenswrapper[4800]: I0122 00:11:48.899665 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.019788 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.026386 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.185549 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.269737 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.363262 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.423603 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.438139 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.449039 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.481318 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.594482 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.601178 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.800120 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 00:11:49 crc kubenswrapper[4800]: I0122 00:11:49.827496 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.085940 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.140029 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.461646 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.502411 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.510508 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.559368 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.860042 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.873726 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 00:11:50 crc kubenswrapper[4800]: I0122 00:11:50.885325 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 00:11:51 crc kubenswrapper[4800]: I0122 00:11:51.017460 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 00:11:51 crc kubenswrapper[4800]: I0122 00:11:51.677691 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55896b6b9d-phdx7"] Jan 22 00:11:51 crc kubenswrapper[4800]: I0122 00:11:51.763416 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 00:11:51 crc kubenswrapper[4800]: I0122 00:11:51.767701 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 00:11:52 crc kubenswrapper[4800]: I0122 00:11:52.207279 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 00:11:52 crc kubenswrapper[4800]: E0122 00:11:52.253965 4800 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 22 00:11:52 crc kubenswrapper[4800]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-55896b6b9d-phdx7_openshift-authentication_5b8263d1-5f1a-4e6e-86b1-d9e622b70d01_0(9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8): error adding pod openshift-authentication_oauth-openshift-55896b6b9d-phdx7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8" Netns:"/var/run/netns/4cb6e976-1c85-418d-aeca-539931f87b82" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-55896b6b9d-phdx7;K8S_POD_INFRA_CONTAINER_ID=9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8;K8S_POD_UID=5b8263d1-5f1a-4e6e-86b1-d9e622b70d01" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-55896b6b9d-phdx7] networking: Multus: [openshift-authentication/oauth-openshift-55896b6b9d-phdx7/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-55896b6b9d-phdx7 in out of cluster comm: pod "oauth-openshift-55896b6b9d-phdx7" not found Jan 22 00:11:52 crc kubenswrapper[4800]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 22 00:11:52 crc kubenswrapper[4800]: > Jan 22 00:11:52 crc kubenswrapper[4800]: E0122 00:11:52.254053 4800 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 22 00:11:52 crc kubenswrapper[4800]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-55896b6b9d-phdx7_openshift-authentication_5b8263d1-5f1a-4e6e-86b1-d9e622b70d01_0(9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8): error adding pod openshift-authentication_oauth-openshift-55896b6b9d-phdx7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8" Netns:"/var/run/netns/4cb6e976-1c85-418d-aeca-539931f87b82" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-55896b6b9d-phdx7;K8S_POD_INFRA_CONTAINER_ID=9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8;K8S_POD_UID=5b8263d1-5f1a-4e6e-86b1-d9e622b70d01" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-55896b6b9d-phdx7] networking: Multus: [openshift-authentication/oauth-openshift-55896b6b9d-phdx7/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-55896b6b9d-phdx7 in out of cluster comm: pod "oauth-openshift-55896b6b9d-phdx7" not found Jan 22 00:11:52 crc kubenswrapper[4800]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 22 00:11:52 crc kubenswrapper[4800]: > pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:52 crc kubenswrapper[4800]: E0122 00:11:52.254082 4800 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 22 00:11:52 crc kubenswrapper[4800]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-55896b6b9d-phdx7_openshift-authentication_5b8263d1-5f1a-4e6e-86b1-d9e622b70d01_0(9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8): error adding pod openshift-authentication_oauth-openshift-55896b6b9d-phdx7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8" Netns:"/var/run/netns/4cb6e976-1c85-418d-aeca-539931f87b82" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-55896b6b9d-phdx7;K8S_POD_INFRA_CONTAINER_ID=9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8;K8S_POD_UID=5b8263d1-5f1a-4e6e-86b1-d9e622b70d01" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-55896b6b9d-phdx7] networking: Multus: [openshift-authentication/oauth-openshift-55896b6b9d-phdx7/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-55896b6b9d-phdx7 in out of cluster comm: pod "oauth-openshift-55896b6b9d-phdx7" not found Jan 22 00:11:52 crc kubenswrapper[4800]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 22 00:11:52 crc kubenswrapper[4800]: > pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:52 crc kubenswrapper[4800]: E0122 00:11:52.254157 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-55896b6b9d-phdx7_openshift-authentication(5b8263d1-5f1a-4e6e-86b1-d9e622b70d01)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-55896b6b9d-phdx7_openshift-authentication(5b8263d1-5f1a-4e6e-86b1-d9e622b70d01)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-55896b6b9d-phdx7_openshift-authentication_5b8263d1-5f1a-4e6e-86b1-d9e622b70d01_0(9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8): error adding pod openshift-authentication_oauth-openshift-55896b6b9d-phdx7 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8\\\" Netns:\\\"/var/run/netns/4cb6e976-1c85-418d-aeca-539931f87b82\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-55896b6b9d-phdx7;K8S_POD_INFRA_CONTAINER_ID=9724860fb611d3e6604f0824b6d0fdb3e6058d2a89e67f41014ccbf30a0914e8;K8S_POD_UID=5b8263d1-5f1a-4e6e-86b1-d9e622b70d01\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-55896b6b9d-phdx7] networking: Multus: [openshift-authentication/oauth-openshift-55896b6b9d-phdx7/5b8263d1-5f1a-4e6e-86b1-d9e622b70d01]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-55896b6b9d-phdx7 in out of cluster comm: pod \\\"oauth-openshift-55896b6b9d-phdx7\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" podUID="5b8263d1-5f1a-4e6e-86b1-d9e622b70d01" Jan 22 00:11:52 crc kubenswrapper[4800]: I0122 00:11:52.274827 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 00:11:52 crc kubenswrapper[4800]: I0122 00:11:52.611794 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:52 crc kubenswrapper[4800]: I0122 00:11:52.612841 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:52 crc kubenswrapper[4800]: I0122 00:11:52.944211 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55896b6b9d-phdx7"] Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.446372 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.542787 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.542855 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.618119 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.618176 4800 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9" exitCode=137 Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.618230 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.618242 4800 scope.go:117] "RemoveContainer" containerID="1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.620021 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" event={"ID":"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01","Type":"ContainerStarted","Data":"0d36adbde24e942d61d3a0c9e9bec3a8214197c769f13a40d3531492ee3d5ba9"} Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.620123 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" event={"ID":"5b8263d1-5f1a-4e6e-86b1-d9e622b70d01","Type":"ContainerStarted","Data":"77057c3052609e9ac4c03ea48085ba4cbc9ce443cc1287bf4e5d3b5a04acc25a"} Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.620233 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.628982 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.639218 4800 scope.go:117] "RemoveContainer" containerID="1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9" Jan 22 00:11:53 crc kubenswrapper[4800]: E0122 00:11:53.639969 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9\": container with ID starting with 1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9 not found: ID does not exist" containerID="1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.640054 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9"} err="failed to get container status \"1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9\": rpc error: code = NotFound desc = could not find container \"1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9\": container with ID starting with 1a3968d2713fd68905b1f531d129dfcfaabb5cf0a8abb383c9071a491e605da9 not found: ID does not exist" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650199 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650257 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650288 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650339 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650381 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650423 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650437 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650548 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.650635 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.651808 4800 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.651838 4800 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.651847 4800 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.651856 4800 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.673360 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.692442 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-55896b6b9d-phdx7" podStartSLOduration=57.69239203 podStartE2EDuration="57.69239203s" podCreationTimestamp="2026-01-22 00:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:11:53.645443719 +0000 UTC m=+289.125746757" watchObservedRunningTime="2026-01-22 00:11:53.69239203 +0000 UTC m=+289.172695068" Jan 22 00:11:53 crc kubenswrapper[4800]: I0122 00:11:53.753025 4800 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 00:11:54 crc kubenswrapper[4800]: I0122 00:11:54.085597 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 00:11:54 crc kubenswrapper[4800]: I0122 00:11:54.828502 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 22 00:12:04 crc kubenswrapper[4800]: I0122 00:12:04.668947 4800 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.439254 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.440158 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerName="controller-manager" containerID="cri-o://e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a" gracePeriod=30 Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.541025 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.541253 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerName="route-controller-manager" containerID="cri-o://dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653" gracePeriod=30 Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.809127 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.853151 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978570 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles\") pod \"5bcffd47-0f44-4095-892e-c3cf4acd975a\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978652 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcxq5\" (UniqueName: \"kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5\") pod \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978710 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca\") pod \"5bcffd47-0f44-4095-892e-c3cf4acd975a\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978775 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert\") pod \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978829 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca\") pod \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978852 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnhj5\" (UniqueName: \"kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5\") pod \"5bcffd47-0f44-4095-892e-c3cf4acd975a\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978878 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config\") pod \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\" (UID: \"9f5ddfd1-20c3-4773-a581-324aad7c3d3f\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.978961 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert\") pod \"5bcffd47-0f44-4095-892e-c3cf4acd975a\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.979002 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config\") pod \"5bcffd47-0f44-4095-892e-c3cf4acd975a\" (UID: \"5bcffd47-0f44-4095-892e-c3cf4acd975a\") " Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.980484 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca" (OuterVolumeSpecName: "client-ca") pod "9f5ddfd1-20c3-4773-a581-324aad7c3d3f" (UID: "9f5ddfd1-20c3-4773-a581-324aad7c3d3f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.980526 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5bcffd47-0f44-4095-892e-c3cf4acd975a" (UID: "5bcffd47-0f44-4095-892e-c3cf4acd975a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.980758 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca" (OuterVolumeSpecName: "client-ca") pod "5bcffd47-0f44-4095-892e-c3cf4acd975a" (UID: "5bcffd47-0f44-4095-892e-c3cf4acd975a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.980988 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config" (OuterVolumeSpecName: "config") pod "9f5ddfd1-20c3-4773-a581-324aad7c3d3f" (UID: "9f5ddfd1-20c3-4773-a581-324aad7c3d3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.981226 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config" (OuterVolumeSpecName: "config") pod "5bcffd47-0f44-4095-892e-c3cf4acd975a" (UID: "5bcffd47-0f44-4095-892e-c3cf4acd975a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.992586 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9f5ddfd1-20c3-4773-a581-324aad7c3d3f" (UID: "9f5ddfd1-20c3-4773-a581-324aad7c3d3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.994660 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5" (OuterVolumeSpecName: "kube-api-access-xnhj5") pod "5bcffd47-0f44-4095-892e-c3cf4acd975a" (UID: "5bcffd47-0f44-4095-892e-c3cf4acd975a"). InnerVolumeSpecName "kube-api-access-xnhj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:12:22 crc kubenswrapper[4800]: I0122 00:12:22.999854 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5" (OuterVolumeSpecName: "kube-api-access-qcxq5") pod "9f5ddfd1-20c3-4773-a581-324aad7c3d3f" (UID: "9f5ddfd1-20c3-4773-a581-324aad7c3d3f"). InnerVolumeSpecName "kube-api-access-qcxq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.000777 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5bcffd47-0f44-4095-892e-c3cf4acd975a" (UID: "5bcffd47-0f44-4095-892e-c3cf4acd975a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019625 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-hbsj2"] Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.019822 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerName="controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019838 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerName="controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.019856 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019862 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.019874 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerName="route-controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019880 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerName="route-controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019980 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerName="route-controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.019997 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.020005 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerName="controller-manager" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.020361 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.047326 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-hbsj2"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.050882 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.051854 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.061821 4800 generic.go:334] "Generic (PLEG): container finished" podID="5bcffd47-0f44-4095-892e-c3cf4acd975a" containerID="e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a" exitCode=0 Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.061994 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.062116 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" event={"ID":"5bcffd47-0f44-4095-892e-c3cf4acd975a","Type":"ContainerDied","Data":"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a"} Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.062161 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lxktv" event={"ID":"5bcffd47-0f44-4095-892e-c3cf4acd975a","Type":"ContainerDied","Data":"f827a6b35faa68019dfc8e802ebec4dafb40459e085d6dcd32eda29d282ac095"} Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.062185 4800 scope.go:117] "RemoveContainer" containerID="e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.065386 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.065964 4800 generic.go:334] "Generic (PLEG): container finished" podID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" containerID="dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653" exitCode=0 Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.066056 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" event={"ID":"9f5ddfd1-20c3-4773-a581-324aad7c3d3f","Type":"ContainerDied","Data":"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653"} Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.066089 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" event={"ID":"9f5ddfd1-20c3-4773-a581-324aad7c3d3f","Type":"ContainerDied","Data":"0dcde9e62663fa1ac328caca41380e23ee362a5f1cc3f58e9f371f7df77d577b"} Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.066139 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080442 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bcffd47-0f44-4095-892e-c3cf4acd975a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080489 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080505 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080519 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcxq5\" (UniqueName: \"kubernetes.io/projected/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-kube-api-access-qcxq5\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080531 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bcffd47-0f44-4095-892e-c3cf4acd975a-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080542 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080552 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnhj5\" (UniqueName: \"kubernetes.io/projected/5bcffd47-0f44-4095-892e-c3cf4acd975a-kube-api-access-xnhj5\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080566 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.080577 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5ddfd1-20c3-4773-a581-324aad7c3d3f-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.090735 4800 scope.go:117] "RemoveContainer" containerID="e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a" Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.098587 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a\": container with ID starting with e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a not found: ID does not exist" containerID="e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.098658 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a"} err="failed to get container status \"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a\": rpc error: code = NotFound desc = could not find container \"e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a\": container with ID starting with e7b74d9a2a6f9facc43ac54cd51751aedbb58551472f148ed1928561f221293a not found: ID does not exist" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.098689 4800 scope.go:117] "RemoveContainer" containerID="dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.130104 4800 scope.go:117] "RemoveContainer" containerID="dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653" Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.139174 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653\": container with ID starting with dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653 not found: ID does not exist" containerID="dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.142097 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.143913 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653"} err="failed to get container status \"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653\": rpc error: code = NotFound desc = could not find container \"dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653\": container with ID starting with dc0797d6121b45c9014d87105876866a2f8b50da071c5db7412058b7a4d1e653 not found: ID does not exist" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.148849 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lxktv"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.161554 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-hbsj2"] Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.161956 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-vvnpl proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" podUID="1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.164844 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh"] Jan 22 00:12:23 crc kubenswrapper[4800]: E0122 00:12:23.165313 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-6rfpb serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" podUID="83352a05-1e97-4aa8-8b26-c71981d7538b" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.171060 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.174065 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-88ls8"] Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.181993 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvnpl\" (UniqueName: \"kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182073 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rfpb\" (UniqueName: \"kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182140 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182166 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182188 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182223 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182279 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182310 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.182337 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284111 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284747 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvnpl\" (UniqueName: \"kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284806 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rfpb\" (UniqueName: \"kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284830 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284847 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284922 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.284944 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.285006 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.285044 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.285880 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.285926 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.285998 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.286381 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.286484 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.287668 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.288077 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.301207 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvnpl\" (UniqueName: \"kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl\") pod \"controller-manager-74ffd4594b-hbsj2\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:23 crc kubenswrapper[4800]: I0122 00:12:23.301283 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rfpb\" (UniqueName: \"kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb\") pod \"route-controller-manager-57df89d5c4-85hxh\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.073111 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.073117 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.080436 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.085461 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.194991 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca\") pod \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195065 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config\") pod \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195097 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles\") pod \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195136 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca\") pod \"83352a05-1e97-4aa8-8b26-c71981d7538b\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195158 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert\") pod \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195190 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rfpb\" (UniqueName: \"kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb\") pod \"83352a05-1e97-4aa8-8b26-c71981d7538b\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195216 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvnpl\" (UniqueName: \"kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl\") pod \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\" (UID: \"1ca9c33e-c7d0-4bc4-afff-eaa95367e86a\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195232 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert\") pod \"83352a05-1e97-4aa8-8b26-c71981d7538b\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195267 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config\") pod \"83352a05-1e97-4aa8-8b26-c71981d7538b\" (UID: \"83352a05-1e97-4aa8-8b26-c71981d7538b\") " Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195655 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca" (OuterVolumeSpecName: "client-ca") pod "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" (UID: "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195722 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca" (OuterVolumeSpecName: "client-ca") pod "83352a05-1e97-4aa8-8b26-c71981d7538b" (UID: "83352a05-1e97-4aa8-8b26-c71981d7538b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.195997 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config" (OuterVolumeSpecName: "config") pod "83352a05-1e97-4aa8-8b26-c71981d7538b" (UID: "83352a05-1e97-4aa8-8b26-c71981d7538b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.196188 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" (UID: "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.196214 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config" (OuterVolumeSpecName: "config") pod "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" (UID: "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.198940 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "83352a05-1e97-4aa8-8b26-c71981d7538b" (UID: "83352a05-1e97-4aa8-8b26-c71981d7538b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.198963 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" (UID: "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.199009 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb" (OuterVolumeSpecName: "kube-api-access-6rfpb") pod "83352a05-1e97-4aa8-8b26-c71981d7538b" (UID: "83352a05-1e97-4aa8-8b26-c71981d7538b"). InnerVolumeSpecName "kube-api-access-6rfpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.205613 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl" (OuterVolumeSpecName: "kube-api-access-vvnpl") pod "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" (UID: "1ca9c33e-c7d0-4bc4-afff-eaa95367e86a"). InnerVolumeSpecName "kube-api-access-vvnpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296414 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296454 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296466 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296476 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296485 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rfpb\" (UniqueName: \"kubernetes.io/projected/83352a05-1e97-4aa8-8b26-c71981d7538b-kube-api-access-6rfpb\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296497 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvnpl\" (UniqueName: \"kubernetes.io/projected/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-kube-api-access-vvnpl\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296505 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83352a05-1e97-4aa8-8b26-c71981d7538b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296513 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83352a05-1e97-4aa8-8b26-c71981d7538b-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.296521 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.826518 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bcffd47-0f44-4095-892e-c3cf4acd975a" path="/var/lib/kubelet/pods/5bcffd47-0f44-4095-892e-c3cf4acd975a/volumes" Jan 22 00:12:24 crc kubenswrapper[4800]: I0122 00:12:24.827305 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f5ddfd1-20c3-4773-a581-324aad7c3d3f" path="/var/lib/kubelet/pods/9f5ddfd1-20c3-4773-a581-324aad7c3d3f/volumes" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.076737 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.076737 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-hbsj2" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.109878 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.110528 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.117259 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.117458 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.117608 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.117911 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.123137 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.123210 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.126850 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-hbsj2"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.128577 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-hbsj2"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.128836 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.141937 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.157934 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.162328 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-85hxh"] Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.309102 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.309164 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.309198 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcnsv\" (UniqueName: \"kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.309263 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.309291 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.409870 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.409969 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.410011 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcnsv\" (UniqueName: \"kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.410041 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.410458 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.411439 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.411978 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.412341 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.425045 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.428269 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcnsv\" (UniqueName: \"kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv\") pod \"controller-manager-7b95dfdc4f-2q6z6\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.431742 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:25 crc kubenswrapper[4800]: I0122 00:12:25.634789 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.082105 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" event={"ID":"acde09bb-7e8f-44f2-b88a-095e00d6ccc2","Type":"ContainerStarted","Data":"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d"} Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.082148 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" event={"ID":"acde09bb-7e8f-44f2-b88a-095e00d6ccc2","Type":"ContainerStarted","Data":"667e537054c0a0d8231f47b0e10d472cf14e54b3a38b0be552c627e1b099f1a5"} Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.083240 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.087711 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.125153 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" podStartSLOduration=3.125129795 podStartE2EDuration="3.125129795s" podCreationTimestamp="2026-01-22 00:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:12:26.102811216 +0000 UTC m=+321.583114254" watchObservedRunningTime="2026-01-22 00:12:26.125129795 +0000 UTC m=+321.605432823" Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.825847 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca9c33e-c7d0-4bc4-afff-eaa95367e86a" path="/var/lib/kubelet/pods/1ca9c33e-c7d0-4bc4-afff-eaa95367e86a/volumes" Jan 22 00:12:26 crc kubenswrapper[4800]: I0122 00:12:26.826559 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83352a05-1e97-4aa8-8b26-c71981d7538b" path="/var/lib/kubelet/pods/83352a05-1e97-4aa8-8b26-c71981d7538b/volumes" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.407815 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.408553 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.411727 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.411987 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.412016 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.412141 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.412201 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.412610 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.425080 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.537960 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.538155 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.538221 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.538306 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8rx6\" (UniqueName: \"kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.640066 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8rx6\" (UniqueName: \"kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.640152 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.640222 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.640238 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.641533 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.641875 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.647327 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.662779 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8rx6\" (UniqueName: \"kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6\") pod \"route-controller-manager-6c595fc99c-l7dkh\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:27 crc kubenswrapper[4800]: I0122 00:12:27.727797 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:28 crc kubenswrapper[4800]: I0122 00:12:28.179551 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:12:28 crc kubenswrapper[4800]: W0122 00:12:28.196174 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ec9b33e_8b1d_453e_815a_cfeba9445d33.slice/crio-d7f6f29cbd1d349eb292299c0c7422f66820db577d7a2c068018f624b3f36d44 WatchSource:0}: Error finding container d7f6f29cbd1d349eb292299c0c7422f66820db577d7a2c068018f624b3f36d44: Status 404 returned error can't find the container with id d7f6f29cbd1d349eb292299c0c7422f66820db577d7a2c068018f624b3f36d44 Jan 22 00:12:29 crc kubenswrapper[4800]: I0122 00:12:29.106722 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" event={"ID":"9ec9b33e-8b1d-453e-815a-cfeba9445d33","Type":"ContainerStarted","Data":"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc"} Jan 22 00:12:29 crc kubenswrapper[4800]: I0122 00:12:29.107432 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:29 crc kubenswrapper[4800]: I0122 00:12:29.107446 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" event={"ID":"9ec9b33e-8b1d-453e-815a-cfeba9445d33","Type":"ContainerStarted","Data":"d7f6f29cbd1d349eb292299c0c7422f66820db577d7a2c068018f624b3f36d44"} Jan 22 00:12:29 crc kubenswrapper[4800]: I0122 00:12:29.112978 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:12:29 crc kubenswrapper[4800]: I0122 00:12:29.124511 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" podStartSLOduration=6.124487775 podStartE2EDuration="6.124487775s" podCreationTimestamp="2026-01-22 00:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:12:29.120094887 +0000 UTC m=+324.600397925" watchObservedRunningTime="2026-01-22 00:12:29.124487775 +0000 UTC m=+324.604790813" Jan 22 00:12:49 crc kubenswrapper[4800]: I0122 00:12:49.452017 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:12:49 crc kubenswrapper[4800]: I0122 00:12:49.452600 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.453777 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.454637 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" podUID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" containerName="route-controller-manager" containerID="cri-o://5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc" gracePeriod=30 Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.892582 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.979499 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config\") pod \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.979581 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert\") pod \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.979603 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca\") pod \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.979651 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8rx6\" (UniqueName: \"kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6\") pod \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\" (UID: \"9ec9b33e-8b1d-453e-815a-cfeba9445d33\") " Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.980379 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca" (OuterVolumeSpecName: "client-ca") pod "9ec9b33e-8b1d-453e-815a-cfeba9445d33" (UID: "9ec9b33e-8b1d-453e-815a-cfeba9445d33"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.980436 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config" (OuterVolumeSpecName: "config") pod "9ec9b33e-8b1d-453e-815a-cfeba9445d33" (UID: "9ec9b33e-8b1d-453e-815a-cfeba9445d33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.984633 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6" (OuterVolumeSpecName: "kube-api-access-n8rx6") pod "9ec9b33e-8b1d-453e-815a-cfeba9445d33" (UID: "9ec9b33e-8b1d-453e-815a-cfeba9445d33"). InnerVolumeSpecName "kube-api-access-n8rx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:13:02 crc kubenswrapper[4800]: I0122 00:13:02.984757 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9ec9b33e-8b1d-453e-815a-cfeba9445d33" (UID: "9ec9b33e-8b1d-453e-815a-cfeba9445d33"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.081500 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ec9b33e-8b1d-453e-815a-cfeba9445d33-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.081549 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.081565 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8rx6\" (UniqueName: \"kubernetes.io/projected/9ec9b33e-8b1d-453e-815a-cfeba9445d33-kube-api-access-n8rx6\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.081578 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec9b33e-8b1d-453e-815a-cfeba9445d33-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.283763 4800 generic.go:334] "Generic (PLEG): container finished" podID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" containerID="5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc" exitCode=0 Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.283828 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.283828 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" event={"ID":"9ec9b33e-8b1d-453e-815a-cfeba9445d33","Type":"ContainerDied","Data":"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc"} Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.283902 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh" event={"ID":"9ec9b33e-8b1d-453e-815a-cfeba9445d33","Type":"ContainerDied","Data":"d7f6f29cbd1d349eb292299c0c7422f66820db577d7a2c068018f624b3f36d44"} Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.283924 4800 scope.go:117] "RemoveContainer" containerID="5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.310129 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.314779 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c595fc99c-l7dkh"] Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.317564 4800 scope.go:117] "RemoveContainer" containerID="5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc" Jan 22 00:13:03 crc kubenswrapper[4800]: E0122 00:13:03.318256 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc\": container with ID starting with 5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc not found: ID does not exist" containerID="5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc" Jan 22 00:13:03 crc kubenswrapper[4800]: I0122 00:13:03.318298 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc"} err="failed to get container status \"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc\": rpc error: code = NotFound desc = could not find container \"5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc\": container with ID starting with 5ae0a990b158dd60cb4a34fe6c8c9eba4fcf9cf7b1d0c73830caf0071fc4cedc not found: ID does not exist" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.432316 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5"] Jan 22 00:13:04 crc kubenswrapper[4800]: E0122 00:13:04.434055 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" containerName="route-controller-manager" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.434147 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" containerName="route-controller-manager" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.434345 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" containerName="route-controller-manager" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.435068 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.443449 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.443698 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.443750 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.443585 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.444201 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.444413 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.462785 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5"] Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.602728 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxnp9\" (UniqueName: \"kubernetes.io/projected/1f788494-6235-4b34-b4f7-df7d71290c43-kube-api-access-mxnp9\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.602798 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f788494-6235-4b34-b4f7-df7d71290c43-serving-cert\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.602822 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-client-ca\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.602860 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-config\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.704016 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxnp9\" (UniqueName: \"kubernetes.io/projected/1f788494-6235-4b34-b4f7-df7d71290c43-kube-api-access-mxnp9\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.704080 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f788494-6235-4b34-b4f7-df7d71290c43-serving-cert\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.704106 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-client-ca\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.704143 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-config\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.705182 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-client-ca\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.705383 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f788494-6235-4b34-b4f7-df7d71290c43-config\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.709491 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f788494-6235-4b34-b4f7-df7d71290c43-serving-cert\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.724189 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxnp9\" (UniqueName: \"kubernetes.io/projected/1f788494-6235-4b34-b4f7-df7d71290c43-kube-api-access-mxnp9\") pod \"route-controller-manager-57df89d5c4-kvxj5\" (UID: \"1f788494-6235-4b34-b4f7-df7d71290c43\") " pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.772955 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.780972 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:04 crc kubenswrapper[4800]: I0122 00:13:04.833091 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ec9b33e-8b1d-453e-815a-cfeba9445d33" path="/var/lib/kubelet/pods/9ec9b33e-8b1d-453e-815a-cfeba9445d33/volumes" Jan 22 00:13:05 crc kubenswrapper[4800]: I0122 00:13:05.190274 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5"] Jan 22 00:13:05 crc kubenswrapper[4800]: I0122 00:13:05.313611 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" event={"ID":"1f788494-6235-4b34-b4f7-df7d71290c43","Type":"ContainerStarted","Data":"a7949ad141d8f25692de348028c84bb2739b094828a6b90c39c517402b40d58f"} Jan 22 00:13:06 crc kubenswrapper[4800]: I0122 00:13:06.322172 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" event={"ID":"1f788494-6235-4b34-b4f7-df7d71290c43","Type":"ContainerStarted","Data":"57b7d0a1097bcf9b3a2c0f4244ac00e54f5dc3b939c3ea310856bfc5f155eb8a"} Jan 22 00:13:06 crc kubenswrapper[4800]: I0122 00:13:06.322771 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:06 crc kubenswrapper[4800]: I0122 00:13:06.328539 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" Jan 22 00:13:06 crc kubenswrapper[4800]: I0122 00:13:06.369247 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-57df89d5c4-kvxj5" podStartSLOduration=4.369220858 podStartE2EDuration="4.369220858s" podCreationTimestamp="2026-01-22 00:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:13:06.348635472 +0000 UTC m=+361.828938510" watchObservedRunningTime="2026-01-22 00:13:06.369220858 +0000 UTC m=+361.849523896" Jan 22 00:13:19 crc kubenswrapper[4800]: I0122 00:13:19.452307 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:13:19 crc kubenswrapper[4800]: I0122 00:13:19.452680 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:13:20 crc kubenswrapper[4800]: I0122 00:13:20.956485 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-svdn5"] Jan 22 00:13:20 crc kubenswrapper[4800]: I0122 00:13:20.957592 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:20 crc kubenswrapper[4800]: I0122 00:13:20.978243 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-svdn5"] Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.021947 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-trusted-ca\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022068 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022096 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m46nz\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-kube-api-access-m46nz\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022146 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-tls\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022162 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022274 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-bound-sa-token\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022425 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.022476 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-certificates\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.048430 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.123976 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-certificates\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124032 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-trusted-ca\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124053 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124071 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m46nz\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-kube-api-access-m46nz\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124102 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-tls\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124121 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124144 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-bound-sa-token\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.124864 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.125451 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-trusted-ca\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.126071 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-certificates\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.136749 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-registry-tls\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.140858 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-bound-sa-token\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.142543 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m46nz\" (UniqueName: \"kubernetes.io/projected/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-kube-api-access-m46nz\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.143146 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-svdn5\" (UID: \"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.276060 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:21 crc kubenswrapper[4800]: I0122 00:13:21.732144 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-svdn5"] Jan 22 00:13:21 crc kubenswrapper[4800]: W0122 00:13:21.748424 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb347cb1_c2a1_4beb_ba35_01c4ca22c6bd.slice/crio-275188bee0f421110838468020ba52a9afe062dc9a9a2512d851055518fa3ac1 WatchSource:0}: Error finding container 275188bee0f421110838468020ba52a9afe062dc9a9a2512d851055518fa3ac1: Status 404 returned error can't find the container with id 275188bee0f421110838468020ba52a9afe062dc9a9a2512d851055518fa3ac1 Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.475628 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.476125 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" podUID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" containerName="controller-manager" containerID="cri-o://92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d" gracePeriod=30 Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.506451 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" event={"ID":"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd","Type":"ContainerStarted","Data":"2010d9b6fc06ac4a6f8aff6f12c9952e555bd23f69cca8f9918d7c9cade8216b"} Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.506867 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" event={"ID":"cb347cb1-c2a1-4beb-ba35-01c4ca22c6bd","Type":"ContainerStarted","Data":"275188bee0f421110838468020ba52a9afe062dc9a9a2512d851055518fa3ac1"} Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.508459 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.536045 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" podStartSLOduration=2.5360264040000002 podStartE2EDuration="2.536026404s" podCreationTimestamp="2026-01-22 00:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:13:22.534675747 +0000 UTC m=+378.014978795" watchObservedRunningTime="2026-01-22 00:13:22.536026404 +0000 UTC m=+378.016329442" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.863705 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.951386 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert\") pod \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.951495 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles\") pod \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.951541 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config\") pod \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.951650 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcnsv\" (UniqueName: \"kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv\") pod \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.951682 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca\") pod \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\" (UID: \"acde09bb-7e8f-44f2-b88a-095e00d6ccc2\") " Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.952732 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca" (OuterVolumeSpecName: "client-ca") pod "acde09bb-7e8f-44f2-b88a-095e00d6ccc2" (UID: "acde09bb-7e8f-44f2-b88a-095e00d6ccc2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.952847 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config" (OuterVolumeSpecName: "config") pod "acde09bb-7e8f-44f2-b88a-095e00d6ccc2" (UID: "acde09bb-7e8f-44f2-b88a-095e00d6ccc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.953156 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "acde09bb-7e8f-44f2-b88a-095e00d6ccc2" (UID: "acde09bb-7e8f-44f2-b88a-095e00d6ccc2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.959207 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv" (OuterVolumeSpecName: "kube-api-access-pcnsv") pod "acde09bb-7e8f-44f2-b88a-095e00d6ccc2" (UID: "acde09bb-7e8f-44f2-b88a-095e00d6ccc2"). InnerVolumeSpecName "kube-api-access-pcnsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:13:22 crc kubenswrapper[4800]: I0122 00:13:22.959266 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "acde09bb-7e8f-44f2-b88a-095e00d6ccc2" (UID: "acde09bb-7e8f-44f2-b88a-095e00d6ccc2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.053874 4800 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.053978 4800 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.053993 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcnsv\" (UniqueName: \"kubernetes.io/projected/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-kube-api-access-pcnsv\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.054037 4800 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.054047 4800 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acde09bb-7e8f-44f2-b88a-095e00d6ccc2-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.514248 4800 generic.go:334] "Generic (PLEG): container finished" podID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" containerID="92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d" exitCode=0 Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.514299 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.514331 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" event={"ID":"acde09bb-7e8f-44f2-b88a-095e00d6ccc2","Type":"ContainerDied","Data":"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d"} Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.514805 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6" event={"ID":"acde09bb-7e8f-44f2-b88a-095e00d6ccc2","Type":"ContainerDied","Data":"667e537054c0a0d8231f47b0e10d472cf14e54b3a38b0be552c627e1b099f1a5"} Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.514846 4800 scope.go:117] "RemoveContainer" containerID="92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.537284 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.538463 4800 scope.go:117] "RemoveContainer" containerID="92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d" Jan 22 00:13:23 crc kubenswrapper[4800]: E0122 00:13:23.538855 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d\": container with ID starting with 92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d not found: ID does not exist" containerID="92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.538922 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d"} err="failed to get container status \"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d\": rpc error: code = NotFound desc = could not find container \"92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d\": container with ID starting with 92689a4b0cf625c37003e03720872d9090ea9fbf571c735ac35d32dcd319393d not found: ID does not exist" Jan 22 00:13:23 crc kubenswrapper[4800]: I0122 00:13:23.539986 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b95dfdc4f-2q6z6"] Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.452413 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-5xf6v"] Jan 22 00:13:24 crc kubenswrapper[4800]: E0122 00:13:24.452659 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" containerName="controller-manager" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.452671 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" containerName="controller-manager" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.452761 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" containerName="controller-manager" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.453179 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.456239 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.456253 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.456366 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.456552 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.456676 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.458256 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.462458 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-5xf6v"] Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.465984 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.573191 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-serving-cert\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.574027 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qpv\" (UniqueName: \"kubernetes.io/projected/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-kube-api-access-n5qpv\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.574161 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.574321 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-client-ca\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.574494 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-config\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.675849 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-config\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.675943 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-serving-cert\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.676000 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qpv\" (UniqueName: \"kubernetes.io/projected/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-kube-api-access-n5qpv\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.676030 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.676057 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-client-ca\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.677516 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-client-ca\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.677676 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-proxy-ca-bundles\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.681008 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-config\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.682781 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-serving-cert\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.696448 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qpv\" (UniqueName: \"kubernetes.io/projected/66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0-kube-api-access-n5qpv\") pod \"controller-manager-74ffd4594b-5xf6v\" (UID: \"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0\") " pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.771135 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:24 crc kubenswrapper[4800]: I0122 00:13:24.825349 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acde09bb-7e8f-44f2-b88a-095e00d6ccc2" path="/var/lib/kubelet/pods/acde09bb-7e8f-44f2-b88a-095e00d6ccc2/volumes" Jan 22 00:13:25 crc kubenswrapper[4800]: W0122 00:13:25.187709 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66daecd7_f1ba_4a4a_9f83_9620fc9ec8c0.slice/crio-377104eed5ebba22cdd7e2f4026b8310d84ac40b406a84d448246ca397f858ea WatchSource:0}: Error finding container 377104eed5ebba22cdd7e2f4026b8310d84ac40b406a84d448246ca397f858ea: Status 404 returned error can't find the container with id 377104eed5ebba22cdd7e2f4026b8310d84ac40b406a84d448246ca397f858ea Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.193106 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74ffd4594b-5xf6v"] Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.527756 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" event={"ID":"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0","Type":"ContainerStarted","Data":"2318a8a8c15c94754e86112956f9fbd3777d75ecf0cf45dcccbd90fd277801b6"} Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.527799 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" event={"ID":"66daecd7-f1ba-4a4a-9f83-9620fc9ec8c0","Type":"ContainerStarted","Data":"377104eed5ebba22cdd7e2f4026b8310d84ac40b406a84d448246ca397f858ea"} Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.528008 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.543220 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" Jan 22 00:13:25 crc kubenswrapper[4800]: I0122 00:13:25.552205 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74ffd4594b-5xf6v" podStartSLOduration=3.552182438 podStartE2EDuration="3.552182438s" podCreationTimestamp="2026-01-22 00:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:13:25.549434753 +0000 UTC m=+381.029737791" watchObservedRunningTime="2026-01-22 00:13:25.552182438 +0000 UTC m=+381.032485476" Jan 22 00:13:41 crc kubenswrapper[4800]: I0122 00:13:41.280146 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-svdn5" Jan 22 00:13:41 crc kubenswrapper[4800]: I0122 00:13:41.334780 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.451418 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.452404 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.452484 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.453420 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.453502 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976" gracePeriod=600 Jan 22 00:13:49 crc kubenswrapper[4800]: E0122 00:13:49.542584 4800 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e02d6ab_3bcb_4fb2_a918_050e2b87ecd1.slice/crio-a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976.scope\": RecentStats: unable to find data in memory cache]" Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.663609 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976" exitCode=0 Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.663674 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976"} Jan 22 00:13:49 crc kubenswrapper[4800]: I0122 00:13:49.664255 4800 scope.go:117] "RemoveContainer" containerID="905c2d2ef5f09eccb2a168bbd5cbacbb6db85a022fe9c7170cd9dfedd24fd2b7" Jan 22 00:13:50 crc kubenswrapper[4800]: I0122 00:13:50.673088 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861"} Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.372145 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" podUID="4b3e309b-172b-4da2-91f5-8be5152cae2e" containerName="registry" containerID="cri-o://756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177" gracePeriod=30 Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.736772 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.778632 4800 generic.go:334] "Generic (PLEG): container finished" podID="4b3e309b-172b-4da2-91f5-8be5152cae2e" containerID="756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177" exitCode=0 Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.778831 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" event={"ID":"4b3e309b-172b-4da2-91f5-8be5152cae2e","Type":"ContainerDied","Data":"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177"} Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.778910 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" event={"ID":"4b3e309b-172b-4da2-91f5-8be5152cae2e","Type":"ContainerDied","Data":"eec5d82aa11ecb1c9de87217a9fd022dccbcb6292efea48048afcfd85d7a6a62"} Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.778963 4800 scope.go:117] "RemoveContainer" containerID="756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.779081 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2qvn4" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.796258 4800 scope.go:117] "RemoveContainer" containerID="756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177" Jan 22 00:14:06 crc kubenswrapper[4800]: E0122 00:14:06.796798 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177\": container with ID starting with 756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177 not found: ID does not exist" containerID="756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.796833 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177"} err="failed to get container status \"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177\": rpc error: code = NotFound desc = could not find container \"756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177\": container with ID starting with 756099bb82b58720055bb1fd2d03f71a4a88cc1bcde79f39b49f578634ca0177 not found: ID does not exist" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883652 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883691 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883725 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f89rt\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883781 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883949 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.883974 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.884002 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.884048 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets\") pod \"4b3e309b-172b-4da2-91f5-8be5152cae2e\" (UID: \"4b3e309b-172b-4da2-91f5-8be5152cae2e\") " Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.884433 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.885059 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.890297 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.890985 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt" (OuterVolumeSpecName: "kube-api-access-f89rt") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "kube-api-access-f89rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.892567 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.893348 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.896473 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.902719 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4b3e309b-172b-4da2-91f5-8be5152cae2e" (UID: "4b3e309b-172b-4da2-91f5-8be5152cae2e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985230 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f89rt\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-kube-api-access-f89rt\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985284 4800 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985307 4800 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b3e309b-172b-4da2-91f5-8be5152cae2e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985324 4800 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4b3e309b-172b-4da2-91f5-8be5152cae2e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985342 4800 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4b3e309b-172b-4da2-91f5-8be5152cae2e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985361 4800 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:06 crc kubenswrapper[4800]: I0122 00:14:06.985549 4800 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4b3e309b-172b-4da2-91f5-8be5152cae2e-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 00:14:07 crc kubenswrapper[4800]: I0122 00:14:07.108964 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:14:07 crc kubenswrapper[4800]: I0122 00:14:07.113125 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2qvn4"] Jan 22 00:14:08 crc kubenswrapper[4800]: I0122 00:14:08.824596 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b3e309b-172b-4da2-91f5-8be5152cae2e" path="/var/lib/kubelet/pods/4b3e309b-172b-4da2-91f5-8be5152cae2e/volumes" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.183034 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp"] Jan 22 00:15:00 crc kubenswrapper[4800]: E0122 00:15:00.184249 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b3e309b-172b-4da2-91f5-8be5152cae2e" containerName="registry" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.184264 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b3e309b-172b-4da2-91f5-8be5152cae2e" containerName="registry" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.184377 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b3e309b-172b-4da2-91f5-8be5152cae2e" containerName="registry" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.184998 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.189123 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.189158 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.196580 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp"] Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.293441 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.293516 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.293545 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pz2j\" (UniqueName: \"kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.395764 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pz2j\" (UniqueName: \"kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.395957 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.396010 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.397399 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.406149 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.421631 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pz2j\" (UniqueName: \"kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j\") pod \"collect-profiles-29484015-fdjdp\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.511172 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:00 crc kubenswrapper[4800]: I0122 00:15:00.725300 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp"] Jan 22 00:15:01 crc kubenswrapper[4800]: I0122 00:15:01.115699 4800 generic.go:334] "Generic (PLEG): container finished" podID="3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" containerID="940c538e81f40ea20719d25e7d6597dc422e3d2090739736d6a32c629f78d15c" exitCode=0 Jan 22 00:15:01 crc kubenswrapper[4800]: I0122 00:15:01.115834 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" event={"ID":"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d","Type":"ContainerDied","Data":"940c538e81f40ea20719d25e7d6597dc422e3d2090739736d6a32c629f78d15c"} Jan 22 00:15:01 crc kubenswrapper[4800]: I0122 00:15:01.116183 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" event={"ID":"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d","Type":"ContainerStarted","Data":"39cc5e8563e7a287f723732f1a44b41c2eb12e71c1c7b5b2ea728f1b558a9c73"} Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.409590 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.523941 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pz2j\" (UniqueName: \"kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j\") pod \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.524013 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume\") pod \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.524079 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume\") pod \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\" (UID: \"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d\") " Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.525304 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume" (OuterVolumeSpecName: "config-volume") pod "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" (UID: "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.531261 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" (UID: "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.532048 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j" (OuterVolumeSpecName: "kube-api-access-6pz2j") pod "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" (UID: "3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d"). InnerVolumeSpecName "kube-api-access-6pz2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.625526 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pz2j\" (UniqueName: \"kubernetes.io/projected/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-kube-api-access-6pz2j\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.625562 4800 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:02 crc kubenswrapper[4800]: I0122 00:15:02.625571 4800 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:03 crc kubenswrapper[4800]: I0122 00:15:03.128649 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" event={"ID":"3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d","Type":"ContainerDied","Data":"39cc5e8563e7a287f723732f1a44b41c2eb12e71c1c7b5b2ea728f1b558a9c73"} Jan 22 00:15:03 crc kubenswrapper[4800]: I0122 00:15:03.128698 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39cc5e8563e7a287f723732f1a44b41c2eb12e71c1c7b5b2ea728f1b558a9c73" Jan 22 00:15:03 crc kubenswrapper[4800]: I0122 00:15:03.128711 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484015-fdjdp" Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.853521 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w4dww"] Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855199 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-controller" containerID="cri-o://b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855255 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="sbdb" containerID="cri-o://e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855378 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-node" containerID="cri-o://728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855377 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855625 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="nbdb" containerID="cri-o://49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855457 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-acl-logging" containerID="cri-o://2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.855643 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="northd" containerID="cri-o://8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" gracePeriod=30 Jan 22 00:15:44 crc kubenswrapper[4800]: I0122 00:15:44.886611 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" containerID="cri-o://ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" gracePeriod=30 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.210283 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/3.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.214214 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovn-acl-logging/0.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.214674 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovn-controller/0.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.215214 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287296 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qwqmf"] Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287598 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="northd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287615 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="northd" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287627 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287633 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287643 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kubecfg-setup" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287649 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kubecfg-setup" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287659 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" containerName="collect-profiles" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287666 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" containerName="collect-profiles" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287679 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287686 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287782 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="sbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287792 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="sbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287802 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="nbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287810 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="nbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287821 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-acl-logging" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287831 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-acl-logging" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287843 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287851 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287860 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287868 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287879 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-node" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287912 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-node" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287924 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287931 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.287941 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.287948 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288077 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288089 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-node" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288101 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288110 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0c48ee-60d3-4c8c-9ae7-9ad9a285b90d" containerName="collect-profiles" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288118 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="sbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288127 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288139 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="nbdb" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288149 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288158 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288165 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="northd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288174 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-acl-logging" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288186 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovn-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.288304 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288313 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.288440 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerName="ovnkube-controller" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.290574 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.360091 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovnkube-controller/3.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.369107 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovn-acl-logging/0.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.369766 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w4dww_46f4c09a-7f56-40ab-b92f-6e01c949a38e/ovn-controller/0.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370411 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370441 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370470 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370478 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370486 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370493 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" exitCode=0 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370502 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" exitCode=143 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370512 4800 generic.go:334] "Generic (PLEG): container finished" podID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" exitCode=143 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370578 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370586 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370638 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370652 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370667 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370678 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370755 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370809 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370735 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370823 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370922 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370932 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370938 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370944 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370950 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370956 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370962 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.370993 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371006 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371015 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371021 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371027 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371033 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371038 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371045 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371095 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371102 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371109 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371127 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371168 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371176 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371183 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371188 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371193 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371198 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371203 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371208 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371273 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371285 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371319 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w4dww" event={"ID":"46f4c09a-7f56-40ab-b92f-6e01c949a38e","Type":"ContainerDied","Data":"76fe4bd599e7342a097e8909937e6c95b31444d2c51f11efe297a01f00745da7"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371336 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371344 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371350 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371357 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371363 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371370 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371849 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371859 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371865 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.371871 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.373414 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/2.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.373979 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/1.log" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.374008 4800 generic.go:334] "Generic (PLEG): container finished" podID="aff2b22c-b87d-47c9-b5a0-6e5772022488" containerID="d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf" exitCode=2 Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.374030 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerDied","Data":"d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.374046 4800 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30"} Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.374428 4800 scope.go:117] "RemoveContainer" containerID="d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.374632 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vfgd4_openshift-multus(aff2b22c-b87d-47c9-b5a0-6e5772022488)\"" pod="openshift-multus/multus-vfgd4" podUID="aff2b22c-b87d-47c9-b5a0-6e5772022488" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.391102 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.399720 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.399991 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.399843 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400024 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400266 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400641 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400407 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400929 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.400992 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401037 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401065 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401070 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gx9m\" (UniqueName: \"kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401117 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401141 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401160 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401181 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401202 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401221 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401288 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401310 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401329 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401350 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401368 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401396 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash\") pod \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\" (UID: \"46f4c09a-7f56-40ab-b92f-6e01c949a38e\") " Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401480 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-config\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401549 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-netns\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401574 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401598 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-systemd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401620 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-node-log\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401641 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-slash\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401681 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-log-socket\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401723 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401744 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-script-lib\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401769 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-etc-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401788 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-bin\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401807 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-netd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401835 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-var-lib-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401903 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-kubelet\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401928 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ce63428-18bb-403b-ad52-687d788fedb4-ovn-node-metrics-cert\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401960 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-env-overrides\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.401980 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402004 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-ovn\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402010 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402026 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gx47\" (UniqueName: \"kubernetes.io/projected/1ce63428-18bb-403b-ad52-687d788fedb4-kube-api-access-7gx47\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402042 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402055 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-systemd-units\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402061 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402099 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402214 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket" (OuterVolumeSpecName: "log-socket") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402246 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402319 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash" (OuterVolumeSpecName: "host-slash") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402427 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log" (OuterVolumeSpecName: "node-log") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402450 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402505 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402532 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402689 4800 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402700 4800 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402708 4800 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402716 4800 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-log-socket\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402723 4800 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402731 4800 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-node-log\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402738 4800 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402745 4800 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-slash\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402753 4800 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402760 4800 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402768 4800 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402775 4800 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.402783 4800 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.403548 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.403589 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.407640 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.408674 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m" (OuterVolumeSpecName: "kube-api-access-2gx9m") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "kube-api-access-2gx9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.411222 4800 scope.go:117] "RemoveContainer" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.418058 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "46f4c09a-7f56-40ab-b92f-6e01c949a38e" (UID: "46f4c09a-7f56-40ab-b92f-6e01c949a38e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.425617 4800 scope.go:117] "RemoveContainer" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.437276 4800 scope.go:117] "RemoveContainer" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.450872 4800 scope.go:117] "RemoveContainer" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.463251 4800 scope.go:117] "RemoveContainer" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.474172 4800 scope.go:117] "RemoveContainer" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.485705 4800 scope.go:117] "RemoveContainer" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.497990 4800 scope.go:117] "RemoveContainer" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504641 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-log-socket\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504706 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504738 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-script-lib\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504765 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-etc-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504789 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-bin\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504798 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-log-socket\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504817 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-netd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504873 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-var-lib-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504935 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-kubelet\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504961 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ce63428-18bb-403b-ad52-687d788fedb4-ovn-node-metrics-cert\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.504991 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-env-overrides\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505021 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505051 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-ovn\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505087 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gx47\" (UniqueName: \"kubernetes.io/projected/1ce63428-18bb-403b-ad52-687d788fedb4-kube-api-access-7gx47\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505115 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-var-lib-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505171 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-bin\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505201 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-systemd-units\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505138 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505206 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505246 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-ovn\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505123 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-systemd-units\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505293 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-kubelet\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505306 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-cni-netd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505336 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-etc-openvswitch\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505746 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-config\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505793 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-netns\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505819 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505848 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-systemd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505873 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-node-log\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505915 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-slash\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505912 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505930 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-run-systemd\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505870 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-run-netns\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.505983 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-node-log\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506017 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ce63428-18bb-403b-ad52-687d788fedb4-host-slash\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506021 4800 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506086 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gx9m\" (UniqueName: \"kubernetes.io/projected/46f4c09a-7f56-40ab-b92f-6e01c949a38e-kube-api-access-2gx9m\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506102 4800 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506115 4800 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f4c09a-7f56-40ab-b92f-6e01c949a38e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506138 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-script-lib\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506132 4800 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506168 4800 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506181 4800 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f4c09a-7f56-40ab-b92f-6e01c949a38e-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506283 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-env-overrides\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.506849 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ce63428-18bb-403b-ad52-687d788fedb4-ovnkube-config\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.511465 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.512058 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.512118 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} err="failed to get container status \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.512165 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.512509 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": container with ID starting with b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd not found: ID does not exist" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.512597 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} err="failed to get container status \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": rpc error: code = NotFound desc = could not find container \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": container with ID starting with b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.512640 4800 scope.go:117] "RemoveContainer" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.513188 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": container with ID starting with e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a not found: ID does not exist" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.513206 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ce63428-18bb-403b-ad52-687d788fedb4-ovn-node-metrics-cert\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.513230 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} err="failed to get container status \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": rpc error: code = NotFound desc = could not find container \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": container with ID starting with e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.513260 4800 scope.go:117] "RemoveContainer" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.513782 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": container with ID starting with 49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8 not found: ID does not exist" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.513833 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} err="failed to get container status \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": rpc error: code = NotFound desc = could not find container \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": container with ID starting with 49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.513864 4800 scope.go:117] "RemoveContainer" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.514267 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": container with ID starting with 8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040 not found: ID does not exist" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.514320 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} err="failed to get container status \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": rpc error: code = NotFound desc = could not find container \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": container with ID starting with 8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.514349 4800 scope.go:117] "RemoveContainer" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.514735 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": container with ID starting with 55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1 not found: ID does not exist" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.514776 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} err="failed to get container status \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": rpc error: code = NotFound desc = could not find container \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": container with ID starting with 55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.514802 4800 scope.go:117] "RemoveContainer" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.515199 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": container with ID starting with 728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48 not found: ID does not exist" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.515246 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} err="failed to get container status \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": rpc error: code = NotFound desc = could not find container \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": container with ID starting with 728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.515276 4800 scope.go:117] "RemoveContainer" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.516156 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": container with ID starting with 2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5 not found: ID does not exist" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.516198 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} err="failed to get container status \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": rpc error: code = NotFound desc = could not find container \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": container with ID starting with 2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.516220 4800 scope.go:117] "RemoveContainer" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.517009 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": container with ID starting with b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9 not found: ID does not exist" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.517060 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} err="failed to get container status \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": rpc error: code = NotFound desc = could not find container \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": container with ID starting with b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.517105 4800 scope.go:117] "RemoveContainer" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: E0122 00:15:45.517783 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": container with ID starting with 73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8 not found: ID does not exist" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.517834 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} err="failed to get container status \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": rpc error: code = NotFound desc = could not find container \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": container with ID starting with 73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.517868 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.518317 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} err="failed to get container status \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.518347 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.518687 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} err="failed to get container status \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": rpc error: code = NotFound desc = could not find container \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": container with ID starting with b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.518727 4800 scope.go:117] "RemoveContainer" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.519191 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} err="failed to get container status \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": rpc error: code = NotFound desc = could not find container \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": container with ID starting with e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.519281 4800 scope.go:117] "RemoveContainer" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.519790 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} err="failed to get container status \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": rpc error: code = NotFound desc = could not find container \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": container with ID starting with 49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.519830 4800 scope.go:117] "RemoveContainer" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.520186 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} err="failed to get container status \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": rpc error: code = NotFound desc = could not find container \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": container with ID starting with 8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.520223 4800 scope.go:117] "RemoveContainer" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.520541 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} err="failed to get container status \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": rpc error: code = NotFound desc = could not find container \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": container with ID starting with 55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.520575 4800 scope.go:117] "RemoveContainer" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.520977 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} err="failed to get container status \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": rpc error: code = NotFound desc = could not find container \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": container with ID starting with 728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.521015 4800 scope.go:117] "RemoveContainer" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.521374 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} err="failed to get container status \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": rpc error: code = NotFound desc = could not find container \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": container with ID starting with 2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.521412 4800 scope.go:117] "RemoveContainer" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.521747 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} err="failed to get container status \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": rpc error: code = NotFound desc = could not find container \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": container with ID starting with b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.521780 4800 scope.go:117] "RemoveContainer" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522164 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} err="failed to get container status \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": rpc error: code = NotFound desc = could not find container \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": container with ID starting with 73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522210 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522615 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} err="failed to get container status \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522639 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522950 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} err="failed to get container status \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": rpc error: code = NotFound desc = could not find container \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": container with ID starting with b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.522968 4800 scope.go:117] "RemoveContainer" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523260 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} err="failed to get container status \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": rpc error: code = NotFound desc = could not find container \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": container with ID starting with e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523277 4800 scope.go:117] "RemoveContainer" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523519 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} err="failed to get container status \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": rpc error: code = NotFound desc = could not find container \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": container with ID starting with 49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523548 4800 scope.go:117] "RemoveContainer" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523922 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} err="failed to get container status \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": rpc error: code = NotFound desc = could not find container \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": container with ID starting with 8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.523949 4800 scope.go:117] "RemoveContainer" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524229 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} err="failed to get container status \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": rpc error: code = NotFound desc = could not find container \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": container with ID starting with 55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524249 4800 scope.go:117] "RemoveContainer" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524560 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} err="failed to get container status \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": rpc error: code = NotFound desc = could not find container \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": container with ID starting with 728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524585 4800 scope.go:117] "RemoveContainer" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524915 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} err="failed to get container status \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": rpc error: code = NotFound desc = could not find container \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": container with ID starting with 2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.524931 4800 scope.go:117] "RemoveContainer" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.525368 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} err="failed to get container status \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": rpc error: code = NotFound desc = could not find container \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": container with ID starting with b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.525468 4800 scope.go:117] "RemoveContainer" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.526067 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} err="failed to get container status \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": rpc error: code = NotFound desc = could not find container \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": container with ID starting with 73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.526099 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.526551 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} err="failed to get container status \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.526674 4800 scope.go:117] "RemoveContainer" containerID="b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.527378 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd"} err="failed to get container status \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": rpc error: code = NotFound desc = could not find container \"b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd\": container with ID starting with b51986e4e70d843ff4eb01f568842c5a05ac5224d574e8d8318f29a191ed00fd not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.527408 4800 scope.go:117] "RemoveContainer" containerID="e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.527770 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a"} err="failed to get container status \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": rpc error: code = NotFound desc = could not find container \"e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a\": container with ID starting with e0ce44388306825c7bc38d8efab6dc0a821ac89d3915399274c2f16899211b3a not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.527827 4800 scope.go:117] "RemoveContainer" containerID="49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.528404 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8"} err="failed to get container status \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": rpc error: code = NotFound desc = could not find container \"49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8\": container with ID starting with 49752ea8c415030e2bf0ac2febbe8009cce3866da4d1fbd0744c63ebe0a8d9f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.528453 4800 scope.go:117] "RemoveContainer" containerID="8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529160 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040"} err="failed to get container status \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": rpc error: code = NotFound desc = could not find container \"8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040\": container with ID starting with 8f691ad78a5a1a677e89af3a9280edd66fbd151395b8f278005a5339978a5040 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529204 4800 scope.go:117] "RemoveContainer" containerID="55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529493 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1"} err="failed to get container status \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": rpc error: code = NotFound desc = could not find container \"55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1\": container with ID starting with 55f0bb9da909bda4c66ebf0055ee59e9c24f499d8bded757e993fa000acb44d1 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529527 4800 scope.go:117] "RemoveContainer" containerID="728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529832 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48"} err="failed to get container status \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": rpc error: code = NotFound desc = could not find container \"728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48\": container with ID starting with 728409af264c41dc9929759808353ac3723b368b57d743a1dc47057555165f48 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529867 4800 scope.go:117] "RemoveContainer" containerID="2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.529986 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gx47\" (UniqueName: \"kubernetes.io/projected/1ce63428-18bb-403b-ad52-687d788fedb4-kube-api-access-7gx47\") pod \"ovnkube-node-qwqmf\" (UID: \"1ce63428-18bb-403b-ad52-687d788fedb4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.530411 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5"} err="failed to get container status \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": rpc error: code = NotFound desc = could not find container \"2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5\": container with ID starting with 2251d81b2d9fa51c45a1ce2a507b00063cc3aec06860b0de53087c9ae53a09d5 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.530534 4800 scope.go:117] "RemoveContainer" containerID="b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.530836 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9"} err="failed to get container status \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": rpc error: code = NotFound desc = could not find container \"b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9\": container with ID starting with b785d44113939abac44700c1f2a8517ae07b228c7ac53084274046593c4376b9 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.530867 4800 scope.go:117] "RemoveContainer" containerID="73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.531186 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8"} err="failed to get container status \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": rpc error: code = NotFound desc = could not find container \"73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8\": container with ID starting with 73e87ce403126eb31776d7ac1d30f6e85b0d94513b66c51ed789522ef40cc0f8 not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.531209 4800 scope.go:117] "RemoveContainer" containerID="ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.531646 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c"} err="failed to get container status \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": rpc error: code = NotFound desc = could not find container \"ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c\": container with ID starting with ba7f98e59416f635dc3af8842913cdac308d4b9f395864d7a7a26f31e0b1982c not found: ID does not exist" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.606077 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.702010 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w4dww"] Jan 22 00:15:45 crc kubenswrapper[4800]: I0122 00:15:45.706352 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w4dww"] Jan 22 00:15:46 crc kubenswrapper[4800]: I0122 00:15:46.381415 4800 generic.go:334] "Generic (PLEG): container finished" podID="1ce63428-18bb-403b-ad52-687d788fedb4" containerID="49f9df66f236ced9bf99cbe6a6fad0797c6bbcd36d7cc8f2810d44ed236e5fe2" exitCode=0 Jan 22 00:15:46 crc kubenswrapper[4800]: I0122 00:15:46.381493 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerDied","Data":"49f9df66f236ced9bf99cbe6a6fad0797c6bbcd36d7cc8f2810d44ed236e5fe2"} Jan 22 00:15:46 crc kubenswrapper[4800]: I0122 00:15:46.381557 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"b58f99362dbe99b2eefeaa06f85c4ca304068f2a12e41785abcce029d1ebc0f0"} Jan 22 00:15:46 crc kubenswrapper[4800]: I0122 00:15:46.824867 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f4c09a-7f56-40ab-b92f-6e01c949a38e" path="/var/lib/kubelet/pods/46f4c09a-7f56-40ab-b92f-6e01c949a38e/volumes" Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391077 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"249e194eeafc07ef000919e2f7c6771a348b285fb539055d4100d9fd15571f66"} Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391533 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"07b84fc5bf81577dca2c278e60aa99af8bcb797fe0eb29c60255d19ee2d1016e"} Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391554 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"11195d1f71d435d20b06f975e99cb1493c1dccae428e5617056cd980ae5ab44d"} Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391570 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"8450dacd7bf2851e1cd55dbec58cc0738e82ec011ff8f134792edf203a1fffc8"} Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391583 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"42aafb3b72ec066f0414393ec0c175d94e75a4b5159b73acf7df30162c29061b"} Jan 22 00:15:47 crc kubenswrapper[4800]: I0122 00:15:47.391595 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"dc4df56e0dcf117e3904dfa647c26d4e8246fcb9cfa14e802b721bec58f79b99"} Jan 22 00:15:49 crc kubenswrapper[4800]: I0122 00:15:49.409547 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"84bb10ce452283f49548f2dc8c48d2a5b3e4c936197637155aaec51f57519e73"} Jan 22 00:15:49 crc kubenswrapper[4800]: I0122 00:15:49.452206 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:15:49 crc kubenswrapper[4800]: I0122 00:15:49.452275 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:15:52 crc kubenswrapper[4800]: I0122 00:15:52.429555 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" event={"ID":"1ce63428-18bb-403b-ad52-687d788fedb4","Type":"ContainerStarted","Data":"8d07d46bd3cb4c787779e66b08ce5b5db7447e678dba6c260127bfcce0ceeec8"} Jan 22 00:15:52 crc kubenswrapper[4800]: I0122 00:15:52.430098 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:52 crc kubenswrapper[4800]: I0122 00:15:52.430113 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:52 crc kubenswrapper[4800]: I0122 00:15:52.461612 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:52 crc kubenswrapper[4800]: I0122 00:15:52.471140 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" podStartSLOduration=7.471113915 podStartE2EDuration="7.471113915s" podCreationTimestamp="2026-01-22 00:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:15:52.466258844 +0000 UTC m=+527.946561892" watchObservedRunningTime="2026-01-22 00:15:52.471113915 +0000 UTC m=+527.951416963" Jan 22 00:15:53 crc kubenswrapper[4800]: I0122 00:15:53.436568 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:53 crc kubenswrapper[4800]: I0122 00:15:53.474652 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:15:57 crc kubenswrapper[4800]: I0122 00:15:57.817937 4800 scope.go:117] "RemoveContainer" containerID="d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf" Jan 22 00:15:57 crc kubenswrapper[4800]: E0122 00:15:57.818603 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vfgd4_openshift-multus(aff2b22c-b87d-47c9-b5a0-6e5772022488)\"" pod="openshift-multus/multus-vfgd4" podUID="aff2b22c-b87d-47c9-b5a0-6e5772022488" Jan 22 00:16:05 crc kubenswrapper[4800]: I0122 00:16:05.128479 4800 scope.go:117] "RemoveContainer" containerID="dc1c5fb08514854fb58063a4a3615cf17b597aeb7ae8255de3db7235943d47ac" Jan 22 00:16:05 crc kubenswrapper[4800]: I0122 00:16:05.162606 4800 scope.go:117] "RemoveContainer" containerID="7e82c8420910989ee50d2b3c160baecdd6265624c31e5f2919d4ec2b9bb9eb30" Jan 22 00:16:05 crc kubenswrapper[4800]: I0122 00:16:05.510734 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/2.log" Jan 22 00:16:11 crc kubenswrapper[4800]: I0122 00:16:11.818826 4800 scope.go:117] "RemoveContainer" containerID="d78580ad4f1d768c60891c7e5bab83f4dca3b0ec8890211aa4889f064cfbaacf" Jan 22 00:16:12 crc kubenswrapper[4800]: I0122 00:16:12.552940 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/2.log" Jan 22 00:16:12 crc kubenswrapper[4800]: I0122 00:16:12.553355 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vfgd4" event={"ID":"aff2b22c-b87d-47c9-b5a0-6e5772022488","Type":"ContainerStarted","Data":"2c052237e5cd452911b6937a1d3b720172803354088c806f0b8fa0a24126a2d0"} Jan 22 00:16:15 crc kubenswrapper[4800]: I0122 00:16:15.627556 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qwqmf" Jan 22 00:16:19 crc kubenswrapper[4800]: I0122 00:16:19.452277 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:16:19 crc kubenswrapper[4800]: I0122 00:16:19.452616 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.452460 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.453076 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.453120 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.453670 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.453713 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861" gracePeriod=600 Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.759967 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861" exitCode=0 Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.760168 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861"} Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.760408 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c"} Jan 22 00:16:49 crc kubenswrapper[4800]: I0122 00:16:49.760438 4800 scope.go:117] "RemoveContainer" containerID="a48bf7316e5d6662911077e8bc2b45277d0fa9f81d321d64c1d27a8fdfe34976" Jan 22 00:16:51 crc kubenswrapper[4800]: I0122 00:16:51.760737 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:16:51 crc kubenswrapper[4800]: I0122 00:16:51.761446 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kckpd" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="registry-server" containerID="cri-o://c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de" gracePeriod=30 Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.114436 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.248769 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzjll\" (UniqueName: \"kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll\") pod \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.248847 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities\") pod \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.248915 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content\") pod \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\" (UID: \"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7\") " Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.250419 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities" (OuterVolumeSpecName: "utilities") pod "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" (UID: "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.255730 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll" (OuterVolumeSpecName: "kube-api-access-bzjll") pod "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" (UID: "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7"). InnerVolumeSpecName "kube-api-access-bzjll". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.285749 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" (UID: "2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.349856 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzjll\" (UniqueName: \"kubernetes.io/projected/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-kube-api-access-bzjll\") on node \"crc\" DevicePath \"\"" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.349929 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.349941 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.780188 4800 generic.go:334] "Generic (PLEG): container finished" podID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerID="c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de" exitCode=0 Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.780242 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerDied","Data":"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de"} Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.780249 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kckpd" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.780281 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kckpd" event={"ID":"2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7","Type":"ContainerDied","Data":"8d31728e6b52f4706e06ec8ffcafa356c10affe6e5921c1e8eb1f2e3bb7c7035"} Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.780305 4800 scope.go:117] "RemoveContainer" containerID="c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.806848 4800 scope.go:117] "RemoveContainer" containerID="6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.809474 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.813617 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kckpd"] Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.825334 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" path="/var/lib/kubelet/pods/2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7/volumes" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.835216 4800 scope.go:117] "RemoveContainer" containerID="f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.857737 4800 scope.go:117] "RemoveContainer" containerID="c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de" Jan 22 00:16:52 crc kubenswrapper[4800]: E0122 00:16:52.860233 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de\": container with ID starting with c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de not found: ID does not exist" containerID="c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.860271 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de"} err="failed to get container status \"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de\": rpc error: code = NotFound desc = could not find container \"c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de\": container with ID starting with c27535ad230a01bba430024741fbb08757365bdee77a6d45e8985347605f59de not found: ID does not exist" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.860294 4800 scope.go:117] "RemoveContainer" containerID="6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed" Jan 22 00:16:52 crc kubenswrapper[4800]: E0122 00:16:52.861614 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed\": container with ID starting with 6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed not found: ID does not exist" containerID="6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.861644 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed"} err="failed to get container status \"6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed\": rpc error: code = NotFound desc = could not find container \"6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed\": container with ID starting with 6e035af445bbd01ddfdba58dfef59a8b25acc1a3197a652d25f56512a030eeed not found: ID does not exist" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.861686 4800 scope.go:117] "RemoveContainer" containerID="f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081" Jan 22 00:16:52 crc kubenswrapper[4800]: E0122 00:16:52.861982 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081\": container with ID starting with f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081 not found: ID does not exist" containerID="f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081" Jan 22 00:16:52 crc kubenswrapper[4800]: I0122 00:16:52.863778 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081"} err="failed to get container status \"f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081\": rpc error: code = NotFound desc = could not find container \"f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081\": container with ID starting with f359f3ec4b5f8ea82c4a32c030158e31b91a88a8d5ca5b3d0ca8cbe7cf53d081 not found: ID does not exist" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.491434 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj"] Jan 22 00:16:55 crc kubenswrapper[4800]: E0122 00:16:55.491946 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="extract-content" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.491960 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="extract-content" Jan 22 00:16:55 crc kubenswrapper[4800]: E0122 00:16:55.491972 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="extract-utilities" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.491978 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="extract-utilities" Jan 22 00:16:55 crc kubenswrapper[4800]: E0122 00:16:55.491988 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="registry-server" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.491994 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="registry-server" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.492084 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f77d02c-3957-4ebe-bdc5-e17d73bf4eb7" containerName="registry-server" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.492754 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.494273 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.533657 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj"] Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.686280 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.686354 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-867np\" (UniqueName: \"kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.686387 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.787859 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.787956 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-867np\" (UniqueName: \"kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.787996 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.788411 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.788443 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:55 crc kubenswrapper[4800]: I0122 00:16:55.814743 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-867np\" (UniqueName: \"kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.106281 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.298561 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj"] Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.801726 4800 generic.go:334] "Generic (PLEG): container finished" podID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerID="b69280f44c852d45a1d7a5691d22ac20adb6415490d8ea75662d97a53db55ffa" exitCode=0 Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.801828 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" event={"ID":"9e04ccf1-c955-4975-acb3-5e92cbab3469","Type":"ContainerDied","Data":"b69280f44c852d45a1d7a5691d22ac20adb6415490d8ea75662d97a53db55ffa"} Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.802142 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" event={"ID":"9e04ccf1-c955-4975-acb3-5e92cbab3469","Type":"ContainerStarted","Data":"9b6bb34c510e89b1fea5ff13e44accda5b89792bd04778275512e96b9ed44c81"} Jan 22 00:16:56 crc kubenswrapper[4800]: I0122 00:16:56.804483 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:16:58 crc kubenswrapper[4800]: I0122 00:16:58.814970 4800 generic.go:334] "Generic (PLEG): container finished" podID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerID="5847f46316bed949c5607db94528a05cebddc4b3c686a9561be5bd4c20697d65" exitCode=0 Jan 22 00:16:58 crc kubenswrapper[4800]: I0122 00:16:58.815095 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" event={"ID":"9e04ccf1-c955-4975-acb3-5e92cbab3469","Type":"ContainerDied","Data":"5847f46316bed949c5607db94528a05cebddc4b3c686a9561be5bd4c20697d65"} Jan 22 00:16:59 crc kubenswrapper[4800]: I0122 00:16:59.833128 4800 generic.go:334] "Generic (PLEG): container finished" podID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerID="cad24710177e4c280f715adb88acabc117e1320a3c537f23238d705ac0ec1bf4" exitCode=0 Jan 22 00:16:59 crc kubenswrapper[4800]: I0122 00:16:59.833257 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" event={"ID":"9e04ccf1-c955-4975-acb3-5e92cbab3469","Type":"ContainerDied","Data":"cad24710177e4c280f715adb88acabc117e1320a3c537f23238d705ac0ec1bf4"} Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.083323 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.263911 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util\") pod \"9e04ccf1-c955-4975-acb3-5e92cbab3469\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.264316 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle\") pod \"9e04ccf1-c955-4975-acb3-5e92cbab3469\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.264517 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-867np\" (UniqueName: \"kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np\") pod \"9e04ccf1-c955-4975-acb3-5e92cbab3469\" (UID: \"9e04ccf1-c955-4975-acb3-5e92cbab3469\") " Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.266340 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle" (OuterVolumeSpecName: "bundle") pod "9e04ccf1-c955-4975-acb3-5e92cbab3469" (UID: "9e04ccf1-c955-4975-acb3-5e92cbab3469"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.270330 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np" (OuterVolumeSpecName: "kube-api-access-867np") pod "9e04ccf1-c955-4975-acb3-5e92cbab3469" (UID: "9e04ccf1-c955-4975-acb3-5e92cbab3469"). InnerVolumeSpecName "kube-api-access-867np". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.277878 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util" (OuterVolumeSpecName: "util") pod "9e04ccf1-c955-4975-acb3-5e92cbab3469" (UID: "9e04ccf1-c955-4975-acb3-5e92cbab3469"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.367923 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-867np\" (UniqueName: \"kubernetes.io/projected/9e04ccf1-c955-4975-acb3-5e92cbab3469-kube-api-access-867np\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.367995 4800 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-util\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.368016 4800 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e04ccf1-c955-4975-acb3-5e92cbab3469-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.847057 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" event={"ID":"9e04ccf1-c955-4975-acb3-5e92cbab3469","Type":"ContainerDied","Data":"9b6bb34c510e89b1fea5ff13e44accda5b89792bd04778275512e96b9ed44c81"} Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.847097 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b6bb34c510e89b1fea5ff13e44accda5b89792bd04778275512e96b9ed44c81" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.847932 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.921918 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg"] Jan 22 00:17:01 crc kubenswrapper[4800]: E0122 00:17:01.923311 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="extract" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.923335 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="extract" Jan 22 00:17:01 crc kubenswrapper[4800]: E0122 00:17:01.923348 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="pull" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.923357 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="pull" Jan 22 00:17:01 crc kubenswrapper[4800]: E0122 00:17:01.923378 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="util" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.923388 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="util" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.924133 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e04ccf1-c955-4975-acb3-5e92cbab3469" containerName="extract" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.929144 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.934755 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.949777 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg"] Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.975617 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.975751 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:01 crc kubenswrapper[4800]: I0122 00:17:01.975841 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpc8j\" (UniqueName: \"kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.077976 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.078106 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.078211 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpc8j\" (UniqueName: \"kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.078748 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.079120 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.098042 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpc8j\" (UniqueName: \"kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.264483 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.470921 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg"] Jan 22 00:17:02 crc kubenswrapper[4800]: W0122 00:17:02.483108 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc2feeb1_c6a6_4b6a_956f_e5fd2f72d74b.slice/crio-faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544 WatchSource:0}: Error finding container faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544: Status 404 returned error can't find the container with id faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544 Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.856196 4800 generic.go:334] "Generic (PLEG): container finished" podID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerID="96a02e7878f95507bc6fe21782a0988bc9be6686ecdaec1673e2f66e18b41d72" exitCode=0 Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.856270 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" event={"ID":"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b","Type":"ContainerDied","Data":"96a02e7878f95507bc6fe21782a0988bc9be6686ecdaec1673e2f66e18b41d72"} Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.856581 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" event={"ID":"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b","Type":"ContainerStarted","Data":"faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544"} Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.895230 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x"] Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.896269 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:02 crc kubenswrapper[4800]: I0122 00:17:02.905705 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x"] Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.092082 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.092167 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.092211 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhrmn\" (UniqueName: \"kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.193602 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.193724 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.193782 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhrmn\" (UniqueName: \"kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.194848 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.194981 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.222347 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhrmn\" (UniqueName: \"kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.520161 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.710110 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x"] Jan 22 00:17:03 crc kubenswrapper[4800]: W0122 00:17:03.717556 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod358633f6_ece5_4055_98a2_25f2805e3e06.slice/crio-501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199 WatchSource:0}: Error finding container 501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199: Status 404 returned error can't find the container with id 501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199 Jan 22 00:17:03 crc kubenswrapper[4800]: I0122 00:17:03.868514 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerStarted","Data":"501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199"} Jan 22 00:17:04 crc kubenswrapper[4800]: I0122 00:17:04.875860 4800 generic.go:334] "Generic (PLEG): container finished" podID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerID="d974d98cf5623f1065af4cf6d4fd1b39b70166ec653912b5a74cbdfc934f7c64" exitCode=0 Jan 22 00:17:04 crc kubenswrapper[4800]: I0122 00:17:04.876004 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" event={"ID":"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b","Type":"ContainerDied","Data":"d974d98cf5623f1065af4cf6d4fd1b39b70166ec653912b5a74cbdfc934f7c64"} Jan 22 00:17:04 crc kubenswrapper[4800]: I0122 00:17:04.887669 4800 generic.go:334] "Generic (PLEG): container finished" podID="358633f6-ece5-4055-98a2-25f2805e3e06" containerID="d497a5b359cc84a462ceaa0acf8174607179b6ae1d85c0afd249f4709b92bd18" exitCode=0 Jan 22 00:17:04 crc kubenswrapper[4800]: I0122 00:17:04.887729 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerDied","Data":"d497a5b359cc84a462ceaa0acf8174607179b6ae1d85c0afd249f4709b92bd18"} Jan 22 00:17:05 crc kubenswrapper[4800]: I0122 00:17:05.219973 4800 scope.go:117] "RemoveContainer" containerID="ec881742b6943307c271c681c9e6725d1cbff1daa6f7a8c7dd5a631594a1ca89" Jan 22 00:17:05 crc kubenswrapper[4800]: I0122 00:17:05.240340 4800 scope.go:117] "RemoveContainer" containerID="8748a52f9f3d3c4e522463c4d3c8f09c3ce7dcd196629a422cbcf2b9b24bab05" Jan 22 00:17:05 crc kubenswrapper[4800]: I0122 00:17:05.894072 4800 generic.go:334] "Generic (PLEG): container finished" podID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerID="2358f486a2a4aa6ada5f80b6aab36ae49120fbb71b4433083b46a761444c78cc" exitCode=0 Jan 22 00:17:05 crc kubenswrapper[4800]: I0122 00:17:05.894164 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" event={"ID":"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b","Type":"ContainerDied","Data":"2358f486a2a4aa6ada5f80b6aab36ae49120fbb71b4433083b46a761444c78cc"} Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.247428 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.255906 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpc8j\" (UniqueName: \"kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j\") pod \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.255955 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util\") pod \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.255998 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle\") pod \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\" (UID: \"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b\") " Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.257064 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle" (OuterVolumeSpecName: "bundle") pod "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" (UID: "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.261989 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j" (OuterVolumeSpecName: "kube-api-access-dpc8j") pod "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" (UID: "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b"). InnerVolumeSpecName "kube-api-access-dpc8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.275370 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util" (OuterVolumeSpecName: "util") pod "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" (UID: "bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.356822 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpc8j\" (UniqueName: \"kubernetes.io/projected/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-kube-api-access-dpc8j\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.356852 4800 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-util\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.356863 4800 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.910850 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" event={"ID":"bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b","Type":"ContainerDied","Data":"faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544"} Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.911278 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faaacc18f42586f0c4fc262a38369da2592df76dabe311edd408ecc73aa90544" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.910899 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg" Jan 22 00:17:07 crc kubenswrapper[4800]: I0122 00:17:07.918628 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerStarted","Data":"72674f0d3de3738a6a70853ca6eb7458b691ae7c1ba32f970ac9a217aa4e44f7"} Jan 22 00:17:08 crc kubenswrapper[4800]: I0122 00:17:08.925112 4800 generic.go:334] "Generic (PLEG): container finished" podID="358633f6-ece5-4055-98a2-25f2805e3e06" containerID="72674f0d3de3738a6a70853ca6eb7458b691ae7c1ba32f970ac9a217aa4e44f7" exitCode=0 Jan 22 00:17:08 crc kubenswrapper[4800]: I0122 00:17:08.925152 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerDied","Data":"72674f0d3de3738a6a70853ca6eb7458b691ae7c1ba32f970ac9a217aa4e44f7"} Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.709635 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk"] Jan 22 00:17:09 crc kubenswrapper[4800]: E0122 00:17:09.710274 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="util" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.710289 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="util" Jan 22 00:17:09 crc kubenswrapper[4800]: E0122 00:17:09.710303 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="extract" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.710310 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="extract" Jan 22 00:17:09 crc kubenswrapper[4800]: E0122 00:17:09.710334 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="pull" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.710343 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="pull" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.710457 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b" containerName="extract" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.711326 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.724351 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk"] Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.889523 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.889661 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.889683 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7k4m\" (UniqueName: \"kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.933910 4800 generic.go:334] "Generic (PLEG): container finished" podID="358633f6-ece5-4055-98a2-25f2805e3e06" containerID="c4209d37d99f2f5968bdf65ed79a48622ca1aa6701b2ffd8ede4adae128c4089" exitCode=0 Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.933948 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerDied","Data":"c4209d37d99f2f5968bdf65ed79a48622ca1aa6701b2ffd8ede4adae128c4089"} Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.990801 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.990853 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7k4m\" (UniqueName: \"kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.990962 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.991432 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:09 crc kubenswrapper[4800]: I0122 00:17:09.991461 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.010003 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7k4m\" (UniqueName: \"kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.023933 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.244831 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk"] Jan 22 00:17:10 crc kubenswrapper[4800]: W0122 00:17:10.259645 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48aa1c61_a36e_4c82_8812_37ceb4f45d59.slice/crio-13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae WatchSource:0}: Error finding container 13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae: Status 404 returned error can't find the container with id 13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.940010 4800 generic.go:334] "Generic (PLEG): container finished" podID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerID="90f1569d4e0ac9f525368e67de42fcea19c58f99a1e6813113c9e1f3db7bb1b6" exitCode=0 Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.940085 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" event={"ID":"48aa1c61-a36e-4c82-8812-37ceb4f45d59","Type":"ContainerDied","Data":"90f1569d4e0ac9f525368e67de42fcea19c58f99a1e6813113c9e1f3db7bb1b6"} Jan 22 00:17:10 crc kubenswrapper[4800]: I0122 00:17:10.940127 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" event={"ID":"48aa1c61-a36e-4c82-8812-37ceb4f45d59","Type":"ContainerStarted","Data":"13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae"} Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.187918 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.309064 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util\") pod \"358633f6-ece5-4055-98a2-25f2805e3e06\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.309179 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle\") pod \"358633f6-ece5-4055-98a2-25f2805e3e06\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.309208 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhrmn\" (UniqueName: \"kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn\") pod \"358633f6-ece5-4055-98a2-25f2805e3e06\" (UID: \"358633f6-ece5-4055-98a2-25f2805e3e06\") " Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.309648 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle" (OuterVolumeSpecName: "bundle") pod "358633f6-ece5-4055-98a2-25f2805e3e06" (UID: "358633f6-ece5-4055-98a2-25f2805e3e06"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.315002 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn" (OuterVolumeSpecName: "kube-api-access-hhrmn") pod "358633f6-ece5-4055-98a2-25f2805e3e06" (UID: "358633f6-ece5-4055-98a2-25f2805e3e06"). InnerVolumeSpecName "kube-api-access-hhrmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.328030 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util" (OuterVolumeSpecName: "util") pod "358633f6-ece5-4055-98a2-25f2805e3e06" (UID: "358633f6-ece5-4055-98a2-25f2805e3e06"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.410903 4800 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.411142 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhrmn\" (UniqueName: \"kubernetes.io/projected/358633f6-ece5-4055-98a2-25f2805e3e06-kube-api-access-hhrmn\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.411199 4800 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/358633f6-ece5-4055-98a2-25f2805e3e06-util\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.476615 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx"] Jan 22 00:17:11 crc kubenswrapper[4800]: E0122 00:17:11.476824 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="pull" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.476837 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="pull" Jan 22 00:17:11 crc kubenswrapper[4800]: E0122 00:17:11.476850 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="extract" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.476860 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="extract" Jan 22 00:17:11 crc kubenswrapper[4800]: E0122 00:17:11.476869 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="util" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.476876 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="util" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.477017 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="358633f6-ece5-4055-98a2-25f2805e3e06" containerName="extract" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.477381 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.488831 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.490500 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.490729 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-w4fk4" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.508014 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.511993 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/68953295-7f82-41aa-a591-a572d99b9210-kube-api-access-kv8z9\") pod \"obo-prometheus-operator-68bc856cb9-wqvbx\" (UID: \"68953295-7f82-41aa-a591-a572d99b9210\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.596748 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.597766 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.600543 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.601226 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-84rmh" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.612787 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/68953295-7f82-41aa-a591-a572d99b9210-kube-api-access-kv8z9\") pod \"obo-prometheus-operator-68bc856cb9-wqvbx\" (UID: \"68953295-7f82-41aa-a591-a572d99b9210\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.612834 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.612880 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.614333 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.614997 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.625166 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.644575 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.645605 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv8z9\" (UniqueName: \"kubernetes.io/projected/68953295-7f82-41aa-a591-a572d99b9210-kube-api-access-kv8z9\") pod \"obo-prometheus-operator-68bc856cb9-wqvbx\" (UID: \"68953295-7f82-41aa-a591-a572d99b9210\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.713683 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.713732 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.713762 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.713805 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.717242 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.722341 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0771650e-3769-4dd8-baff-e6fc99f841a4-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x\" (UID: \"0771650e-3769-4dd8-baff-e6fc99f841a4\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.787013 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-zhrc2"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.787827 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.789747 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.790085 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2f6tl" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.797339 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.807940 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-zhrc2"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.815635 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.816007 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.816042 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcw2j\" (UniqueName: \"kubernetes.io/projected/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-kube-api-access-fcw2j\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.816199 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.819599 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.822310 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb96f099-5861-45a5-89e7-2ab5c80a90cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq\" (UID: \"fb96f099-5861-45a5-89e7-2ab5c80a90cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.913005 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-p9zm8"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.914447 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.914482 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.916865 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c78d4-97fb-48c5-931d-8477f4876451-openshift-service-ca\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.917068 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.917155 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcw2j\" (UniqueName: \"kubernetes.io/projected/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-kube-api-access-fcw2j\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.917208 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvz5c\" (UniqueName: \"kubernetes.io/projected/2b7c78d4-97fb-48c5-931d-8477f4876451-kube-api-access-qvz5c\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.921181 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.922067 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-vwgqh" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.931460 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.940066 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-p9zm8"] Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.947689 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcw2j\" (UniqueName: \"kubernetes.io/projected/d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3-kube-api-access-fcw2j\") pod \"observability-operator-59bdc8b94-zhrc2\" (UID: \"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3\") " pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.965878 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" event={"ID":"358633f6-ece5-4055-98a2-25f2805e3e06","Type":"ContainerDied","Data":"501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199"} Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.965935 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501a833cf6cdf45b150b25c5d10d56fef2992f97f164964571b8a7de0de09199" Jan 22 00:17:11 crc kubenswrapper[4800]: I0122 00:17:11.966005 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.020716 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvz5c\" (UniqueName: \"kubernetes.io/projected/2b7c78d4-97fb-48c5-931d-8477f4876451-kube-api-access-qvz5c\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.020768 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c78d4-97fb-48c5-931d-8477f4876451-openshift-service-ca\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.021606 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c78d4-97fb-48c5-931d-8477f4876451-openshift-service-ca\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.046709 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvz5c\" (UniqueName: \"kubernetes.io/projected/2b7c78d4-97fb-48c5-931d-8477f4876451-kube-api-access-qvz5c\") pod \"perses-operator-5bf474d74f-p9zm8\" (UID: \"2b7c78d4-97fb-48c5-931d-8477f4876451\") " pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.090745 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx"] Jan 22 00:17:12 crc kubenswrapper[4800]: W0122 00:17:12.102306 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68953295_7f82_41aa_a591_a572d99b9210.slice/crio-0bcf078024a49b52e96ad8b63ecfdc3161c864d729f9a3eaacc7495fc8c298df WatchSource:0}: Error finding container 0bcf078024a49b52e96ad8b63ecfdc3161c864d729f9a3eaacc7495fc8c298df: Status 404 returned error can't find the container with id 0bcf078024a49b52e96ad8b63ecfdc3161c864d729f9a3eaacc7495fc8c298df Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.104571 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.225682 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq"] Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.263441 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.345234 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x"] Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.542573 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-zhrc2"] Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.595567 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-p9zm8"] Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.982047 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" event={"ID":"2b7c78d4-97fb-48c5-931d-8477f4876451","Type":"ContainerStarted","Data":"13fe4f79090886ef0b6d11167bb710ae74be14316198113566221e7c0a17e2e8"} Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.986764 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" event={"ID":"68953295-7f82-41aa-a591-a572d99b9210","Type":"ContainerStarted","Data":"0bcf078024a49b52e96ad8b63ecfdc3161c864d729f9a3eaacc7495fc8c298df"} Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.988081 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" event={"ID":"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3","Type":"ContainerStarted","Data":"a60ecf42b5eebfb871638b7626d9472f63e9bc9b0462ebbeb16af62a8b3fb98f"} Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.988840 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" event={"ID":"fb96f099-5861-45a5-89e7-2ab5c80a90cb","Type":"ContainerStarted","Data":"b29b1127d4e0b42f7d792a139959809f5feab24b72141a93ec78bcde7c4fb1f7"} Jan 22 00:17:12 crc kubenswrapper[4800]: I0122 00:17:12.991451 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" event={"ID":"0771650e-3769-4dd8-baff-e6fc99f841a4","Type":"ContainerStarted","Data":"87c41a6c2e83a1d55a5ce4cb17abc5ccb6b50e0348ac9dec0fa27ecfd4d2324d"} Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.501651 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-m47bj"] Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.502677 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.504791 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.505162 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-cpxrj" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.509949 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.520617 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-m47bj"] Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.551064 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhx2q\" (UniqueName: \"kubernetes.io/projected/e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3-kube-api-access-xhx2q\") pod \"interconnect-operator-5bb49f789d-m47bj\" (UID: \"e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3\") " pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.652147 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhx2q\" (UniqueName: \"kubernetes.io/projected/e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3-kube-api-access-xhx2q\") pod \"interconnect-operator-5bb49f789d-m47bj\" (UID: \"e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3\") " pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.709034 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhx2q\" (UniqueName: \"kubernetes.io/projected/e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3-kube-api-access-xhx2q\") pod \"interconnect-operator-5bb49f789d-m47bj\" (UID: \"e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3\") " pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" Jan 22 00:17:18 crc kubenswrapper[4800]: I0122 00:17:18.821934 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.382874 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-545b84bc9b-ln8zq"] Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.384276 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.390997 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-pw9c5" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.397063 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-545b84bc9b-ln8zq"] Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.397302 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.506220 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-apiservice-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.506305 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-webhook-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.506360 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdzp2\" (UniqueName: \"kubernetes.io/projected/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-kube-api-access-zdzp2\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.607402 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-apiservice-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.607498 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-webhook-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.607538 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdzp2\" (UniqueName: \"kubernetes.io/projected/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-kube-api-access-zdzp2\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.613644 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-apiservice-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.617510 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-webhook-cert\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.622959 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdzp2\" (UniqueName: \"kubernetes.io/projected/a247bbf3-14dc-4b0a-8ae2-27272b6b6f16-kube-api-access-zdzp2\") pod \"elastic-operator-545b84bc9b-ln8zq\" (UID: \"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16\") " pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:22 crc kubenswrapper[4800]: I0122 00:17:22.708103 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" Jan 22 00:17:23 crc kubenswrapper[4800]: I0122 00:17:23.155093 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-545b84bc9b-ln8zq"] Jan 22 00:17:23 crc kubenswrapper[4800]: I0122 00:17:23.206693 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-m47bj"] Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.080358 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" event={"ID":"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16","Type":"ContainerStarted","Data":"6c2e9c3f90302e051b191137408006756397b60c9a3340d06240ac0634df6ca2"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.082438 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" event={"ID":"2b7c78d4-97fb-48c5-931d-8477f4876451","Type":"ContainerStarted","Data":"6e51ffe542c83441bc19dcad33a5e451234164ad1b4fe463a79601aa695c10b9"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.083765 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.091162 4800 generic.go:334] "Generic (PLEG): container finished" podID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerID="ce8b8d2591aa368205aa340bf2855335606691c4b60f9e1f0fddfdea5dc048a4" exitCode=0 Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.091230 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" event={"ID":"48aa1c61-a36e-4c82-8812-37ceb4f45d59","Type":"ContainerDied","Data":"ce8b8d2591aa368205aa340bf2855335606691c4b60f9e1f0fddfdea5dc048a4"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.092754 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" event={"ID":"e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3","Type":"ContainerStarted","Data":"ae0bc1436d287cf34ece39588d0bd7135c1e3317fca5aa3d87e491c76267958b"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.095784 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" event={"ID":"68953295-7f82-41aa-a591-a572d99b9210","Type":"ContainerStarted","Data":"7d13e9e55cf73752650689cf5f8075e4bab273d0515b8e5cad34ec3146e0c3e2"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.102790 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" podStartSLOduration=2.779748414 podStartE2EDuration="13.102769652s" podCreationTimestamp="2026-01-22 00:17:11 +0000 UTC" firstStartedPulling="2026-01-22 00:17:12.611548041 +0000 UTC m=+608.091851079" lastFinishedPulling="2026-01-22 00:17:22.934569279 +0000 UTC m=+618.414872317" observedRunningTime="2026-01-22 00:17:24.097028538 +0000 UTC m=+619.577331586" watchObservedRunningTime="2026-01-22 00:17:24.102769652 +0000 UTC m=+619.583072690" Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.105470 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" event={"ID":"fb96f099-5861-45a5-89e7-2ab5c80a90cb","Type":"ContainerStarted","Data":"ae87b26b2c1c4fc1ae8a4ebf96db6d083c66e3881f220c105695669833026457"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.107672 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" event={"ID":"0771650e-3769-4dd8-baff-e6fc99f841a4","Type":"ContainerStarted","Data":"1bab0beb23f14fc346e6f995030ca26d789a08b1f797a4b85968ebf80af17588"} Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.137442 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-wqvbx" podStartSLOduration=2.340451913 podStartE2EDuration="13.137425763s" podCreationTimestamp="2026-01-22 00:17:11 +0000 UTC" firstStartedPulling="2026-01-22 00:17:12.10712708 +0000 UTC m=+607.587430128" lastFinishedPulling="2026-01-22 00:17:22.90410094 +0000 UTC m=+618.384403978" observedRunningTime="2026-01-22 00:17:24.13506144 +0000 UTC m=+619.615364488" watchObservedRunningTime="2026-01-22 00:17:24.137425763 +0000 UTC m=+619.617728801" Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.211350 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x" podStartSLOduration=2.657943313 podStartE2EDuration="13.211324298s" podCreationTimestamp="2026-01-22 00:17:11 +0000 UTC" firstStartedPulling="2026-01-22 00:17:12.377958406 +0000 UTC m=+607.858261444" lastFinishedPulling="2026-01-22 00:17:22.931339391 +0000 UTC m=+618.411642429" observedRunningTime="2026-01-22 00:17:24.157973485 +0000 UTC m=+619.638276513" watchObservedRunningTime="2026-01-22 00:17:24.211324298 +0000 UTC m=+619.691627336" Jan 22 00:17:24 crc kubenswrapper[4800]: I0122 00:17:24.845696 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq" podStartSLOduration=3.2086913089999998 podStartE2EDuration="13.845677101s" podCreationTimestamp="2026-01-22 00:17:11 +0000 UTC" firstStartedPulling="2026-01-22 00:17:12.267061237 +0000 UTC m=+607.747364275" lastFinishedPulling="2026-01-22 00:17:22.904047029 +0000 UTC m=+618.384350067" observedRunningTime="2026-01-22 00:17:24.215934862 +0000 UTC m=+619.696237920" watchObservedRunningTime="2026-01-22 00:17:24.845677101 +0000 UTC m=+620.325980129" Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.144176 4800 generic.go:334] "Generic (PLEG): container finished" podID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerID="5d65d730feca614b1a3d0caac36bb472f3a5fe41c3fe7ca1ce3a76e4f089c0c9" exitCode=0 Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.144367 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" event={"ID":"48aa1c61-a36e-4c82-8812-37ceb4f45d59","Type":"ContainerDied","Data":"5d65d730feca614b1a3d0caac36bb472f3a5fe41c3fe7ca1ce3a76e4f089c0c9"} Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.147429 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" event={"ID":"d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3","Type":"ContainerStarted","Data":"5c28a4eb0328d5e13c832a6bac41d98f6cc4fcfb5d73d1f8826064f1e162145c"} Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.147674 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.155451 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" Jan 22 00:17:27 crc kubenswrapper[4800]: I0122 00:17:27.212087 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-zhrc2" podStartSLOduration=2.343336381 podStartE2EDuration="16.212067463s" podCreationTimestamp="2026-01-22 00:17:11 +0000 UTC" firstStartedPulling="2026-01-22 00:17:12.557237552 +0000 UTC m=+608.037540590" lastFinishedPulling="2026-01-22 00:17:26.425968634 +0000 UTC m=+621.906271672" observedRunningTime="2026-01-22 00:17:27.209388031 +0000 UTC m=+622.689691069" watchObservedRunningTime="2026-01-22 00:17:27.212067463 +0000 UTC m=+622.692370501" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.071579 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.104503 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util\") pod \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.104546 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle\") pod \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.104584 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7k4m\" (UniqueName: \"kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m\") pod \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\" (UID: \"48aa1c61-a36e-4c82-8812-37ceb4f45d59\") " Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.106378 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle" (OuterVolumeSpecName: "bundle") pod "48aa1c61-a36e-4c82-8812-37ceb4f45d59" (UID: "48aa1c61-a36e-4c82-8812-37ceb4f45d59"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.113227 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m" (OuterVolumeSpecName: "kube-api-access-k7k4m") pod "48aa1c61-a36e-4c82-8812-37ceb4f45d59" (UID: "48aa1c61-a36e-4c82-8812-37ceb4f45d59"). InnerVolumeSpecName "kube-api-access-k7k4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.115196 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util" (OuterVolumeSpecName: "util") pod "48aa1c61-a36e-4c82-8812-37ceb4f45d59" (UID: "48aa1c61-a36e-4c82-8812-37ceb4f45d59"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.170058 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.172111 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk" event={"ID":"48aa1c61-a36e-4c82-8812-37ceb4f45d59","Type":"ContainerDied","Data":"13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae"} Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.172181 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13f401a06bd8c322d1d1d39a11d9c9e61a66580ac0555c90a885c668cf46b1ae" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.206529 4800 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-util\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.206575 4800 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/48aa1c61-a36e-4c82-8812-37ceb4f45d59-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:29 crc kubenswrapper[4800]: I0122 00:17:29.206592 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7k4m\" (UniqueName: \"kubernetes.io/projected/48aa1c61-a36e-4c82-8812-37ceb4f45d59-kube-api-access-k7k4m\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:30 crc kubenswrapper[4800]: I0122 00:17:30.178483 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" event={"ID":"a247bbf3-14dc-4b0a-8ae2-27272b6b6f16","Type":"ContainerStarted","Data":"a02e468436e244eb76d2f099cd14c8190faeb004ac680a7799b89a1fa905df49"} Jan 22 00:17:30 crc kubenswrapper[4800]: I0122 00:17:30.199783 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-545b84bc9b-ln8zq" podStartSLOduration=2.303163499 podStartE2EDuration="8.19975364s" podCreationTimestamp="2026-01-22 00:17:22 +0000 UTC" firstStartedPulling="2026-01-22 00:17:23.178744438 +0000 UTC m=+618.659047476" lastFinishedPulling="2026-01-22 00:17:29.075334579 +0000 UTC m=+624.555637617" observedRunningTime="2026-01-22 00:17:30.199714708 +0000 UTC m=+625.680017746" watchObservedRunningTime="2026-01-22 00:17:30.19975364 +0000 UTC m=+625.680056688" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.271773 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-p9zm8" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.961158 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 00:17:32 crc kubenswrapper[4800]: E0122 00:17:32.961413 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="extract" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.961426 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="extract" Jan 22 00:17:32 crc kubenswrapper[4800]: E0122 00:17:32.961433 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="util" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.961439 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="util" Jan 22 00:17:32 crc kubenswrapper[4800]: E0122 00:17:32.961452 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="pull" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.961459 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="pull" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.961563 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="48aa1c61-a36e-4c82-8812-37ceb4f45d59" containerName="extract" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.962346 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.966349 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.966535 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.966536 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.967703 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.967936 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.968052 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-twqtb" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.968156 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.968278 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.968619 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 22 00:17:32 crc kubenswrapper[4800]: I0122 00:17:32.994077 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088648 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088719 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088738 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088795 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088857 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088878 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088918 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.088964 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089026 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089059 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089087 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9791f1db-7616-4146-8651-2dc03b95fdc7-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089105 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089120 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089144 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.089161 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190110 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190173 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190212 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190241 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190273 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190304 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190343 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9791f1db-7616-4146-8651-2dc03b95fdc7-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190369 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190389 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190410 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190432 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190492 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190518 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190536 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.190557 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.191404 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.192475 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.192536 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.192792 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.193295 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.196962 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.197377 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.197643 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.197977 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9791f1db-7616-4146-8651-2dc03b95fdc7-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.199022 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.199408 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.199653 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.203318 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.203541 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9791f1db-7616-4146-8651-2dc03b95fdc7-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.217662 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9791f1db-7616-4146-8651-2dc03b95fdc7-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9791f1db-7616-4146-8651-2dc03b95fdc7\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:33 crc kubenswrapper[4800]: I0122 00:17:33.289530 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:17:35 crc kubenswrapper[4800]: I0122 00:17:35.337630 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 00:17:36 crc kubenswrapper[4800]: I0122 00:17:36.239263 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" event={"ID":"e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3","Type":"ContainerStarted","Data":"b7089e662e93895d30bb1945576c6221bb95ae62ed477116c2a18369ac142238"} Jan 22 00:17:36 crc kubenswrapper[4800]: I0122 00:17:36.240973 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9791f1db-7616-4146-8651-2dc03b95fdc7","Type":"ContainerStarted","Data":"93f3a54ed95ee4451bac9c02a76416946d2c5401cf6a3672977b61666ae5cd69"} Jan 22 00:17:36 crc kubenswrapper[4800]: I0122 00:17:36.254635 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-m47bj" podStartSLOduration=6.418380803 podStartE2EDuration="18.254608024s" podCreationTimestamp="2026-01-22 00:17:18 +0000 UTC" firstStartedPulling="2026-01-22 00:17:23.216346699 +0000 UTC m=+618.696649737" lastFinishedPulling="2026-01-22 00:17:35.05257392 +0000 UTC m=+630.532876958" observedRunningTime="2026-01-22 00:17:36.25224481 +0000 UTC m=+631.732547868" watchObservedRunningTime="2026-01-22 00:17:36.254608024 +0000 UTC m=+631.734911062" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.463558 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.464854 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.467473 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-t9bnc" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.467957 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.468058 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.468103 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.494612 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640479 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pf2x\" (UniqueName: \"kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640545 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640565 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640583 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640604 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640628 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640661 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640697 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640717 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640738 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640765 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.640800 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742717 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742789 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742816 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742835 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742852 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742875 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742915 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pf2x\" (UniqueName: \"kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742939 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742959 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.742979 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.743000 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.743025 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.743461 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.744350 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.744825 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.744871 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.745059 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.745360 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.745406 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.746047 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.747197 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.753401 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.755178 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.769505 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pf2x\" (UniqueName: \"kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x\") pod \"service-telemetry-operator-1-build\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:41 crc kubenswrapper[4800]: I0122 00:17:41.785654 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:43 crc kubenswrapper[4800]: I0122 00:17:43.691266 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:43 crc kubenswrapper[4800]: W0122 00:17:43.702431 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b5414b4_518d_4c6b_a9cc_f9807a6fe609.slice/crio-f225800aff9a81e106fb0c9246f852f5a6da99503a794ac4569e1ced13d70212 WatchSource:0}: Error finding container f225800aff9a81e106fb0c9246f852f5a6da99503a794ac4569e1ced13d70212: Status 404 returned error can't find the container with id f225800aff9a81e106fb0c9246f852f5a6da99503a794ac4569e1ced13d70212 Jan 22 00:17:44 crc kubenswrapper[4800]: I0122 00:17:44.303549 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3b5414b4-518d-4c6b-a9cc-f9807a6fe609","Type":"ContainerStarted","Data":"f225800aff9a81e106fb0c9246f852f5a6da99503a794ac4569e1ced13d70212"} Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.321328 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw"] Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.322785 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.327137 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.327527 4800 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-z9986" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.327146 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.337799 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw"] Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.455971 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/15b2310b-fa78-47b6-a32b-2c3c3984554d-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.456081 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbssl\" (UniqueName: \"kubernetes.io/projected/15b2310b-fa78-47b6-a32b-2c3c3984554d-kube-api-access-rbssl\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.557065 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/15b2310b-fa78-47b6-a32b-2c3c3984554d-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.557141 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbssl\" (UniqueName: \"kubernetes.io/projected/15b2310b-fa78-47b6-a32b-2c3c3984554d-kube-api-access-rbssl\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.557585 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/15b2310b-fa78-47b6-a32b-2c3c3984554d-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.594579 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbssl\" (UniqueName: \"kubernetes.io/projected/15b2310b-fa78-47b6-a32b-2c3c3984554d-kube-api-access-rbssl\") pod \"cert-manager-operator-controller-manager-5446d6888b-rdnvw\" (UID: \"15b2310b-fa78-47b6-a32b-2c3c3984554d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:47 crc kubenswrapper[4800]: I0122 00:17:47.646352 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" Jan 22 00:17:51 crc kubenswrapper[4800]: I0122 00:17:51.877683 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.520230 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.521741 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.524339 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.524565 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.524817 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.534525 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660160 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660225 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660278 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660302 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660322 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660356 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660382 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660403 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6n8r\" (UniqueName: \"kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660453 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660474 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660499 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.660522 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.761867 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.761934 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.761968 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.761989 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762022 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762037 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762054 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762080 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762102 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762115 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6n8r\" (UniqueName: \"kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762150 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762163 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762528 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.762710 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763073 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763261 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763279 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763627 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763927 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763981 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.763992 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.768266 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.776099 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.779263 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6n8r\" (UniqueName: \"kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r\") pod \"service-telemetry-operator-2-build\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:53 crc kubenswrapper[4800]: I0122 00:17:53.848681 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.288858 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:17:54 crc kubenswrapper[4800]: W0122 00:17:54.295398 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode63289fb_6cba_4893_8e38_ac34600fcd4d.slice/crio-066d77bede158a7763abe903bfbe7c5493be8434015592eb147aa3b362b8948b WatchSource:0}: Error finding container 066d77bede158a7763abe903bfbe7c5493be8434015592eb147aa3b362b8948b: Status 404 returned error can't find the container with id 066d77bede158a7763abe903bfbe7c5493be8434015592eb147aa3b362b8948b Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.316055 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw"] Jan 22 00:17:54 crc kubenswrapper[4800]: W0122 00:17:54.329873 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15b2310b_fa78_47b6_a32b_2c3c3984554d.slice/crio-a16324b3184c4b50bd7fab31238c4c4ff804c14302d4bf3f1111e3fe316a8ab7 WatchSource:0}: Error finding container a16324b3184c4b50bd7fab31238c4c4ff804c14302d4bf3f1111e3fe316a8ab7: Status 404 returned error can't find the container with id a16324b3184c4b50bd7fab31238c4c4ff804c14302d4bf3f1111e3fe316a8ab7 Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.420342 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e63289fb-6cba-4893-8e38-ac34600fcd4d","Type":"ContainerStarted","Data":"066d77bede158a7763abe903bfbe7c5493be8434015592eb147aa3b362b8948b"} Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.421228 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" event={"ID":"15b2310b-fa78-47b6-a32b-2c3c3984554d","Type":"ContainerStarted","Data":"a16324b3184c4b50bd7fab31238c4c4ff804c14302d4bf3f1111e3fe316a8ab7"} Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.422651 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3b5414b4-518d-4c6b-a9cc-f9807a6fe609","Type":"ContainerStarted","Data":"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d"} Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.422826 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" containerName="manage-dockerfile" containerID="cri-o://72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d" gracePeriod=30 Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.425132 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9791f1db-7616-4146-8651-2dc03b95fdc7","Type":"ContainerStarted","Data":"d45606e2750122f5757f1db7e3dda6bedb703c4e76f4c54534ab86cff6a991bd"} Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.670650 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.705881 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.748801 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_3b5414b4-518d-4c6b-a9cc-f9807a6fe609/manage-dockerfile/0.log" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.748936 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896355 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896739 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896774 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pf2x\" (UniqueName: \"kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896837 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896859 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896921 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896959 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896989 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.897055 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.897084 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.897113 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.897140 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles\") pod \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\" (UID: \"3b5414b4-518d-4c6b-a9cc-f9807a6fe609\") " Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.896570 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.897849 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898095 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898324 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898668 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898760 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898777 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.898963 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.899204 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.903645 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x" (OuterVolumeSpecName: "kube-api-access-2pf2x") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "kube-api-access-2pf2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.907071 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-pull") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "builder-dockercfg-t9bnc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.914043 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-push") pod "3b5414b4-518d-4c6b-a9cc-f9807a6fe609" (UID: "3b5414b4-518d-4c6b-a9cc-f9807a6fe609"). InnerVolumeSpecName "builder-dockercfg-t9bnc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999330 4800 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999370 4800 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999382 4800 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999393 4800 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999401 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999409 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pf2x\" (UniqueName: \"kubernetes.io/projected/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-kube-api-access-2pf2x\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999417 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-push\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999425 4800 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999433 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999442 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-builder-dockercfg-t9bnc-pull\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999449 4800 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:54 crc kubenswrapper[4800]: I0122 00:17:54.999457 4800 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3b5414b4-518d-4c6b-a9cc-f9807a6fe609-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432012 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_3b5414b4-518d-4c6b-a9cc-f9807a6fe609/manage-dockerfile/0.log" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432060 4800 generic.go:334] "Generic (PLEG): container finished" podID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" containerID="72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d" exitCode=2 Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432124 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3b5414b4-518d-4c6b-a9cc-f9807a6fe609","Type":"ContainerDied","Data":"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d"} Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432156 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"3b5414b4-518d-4c6b-a9cc-f9807a6fe609","Type":"ContainerDied","Data":"f225800aff9a81e106fb0c9246f852f5a6da99503a794ac4569e1ced13d70212"} Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432174 4800 scope.go:117] "RemoveContainer" containerID="72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.432200 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.434122 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e63289fb-6cba-4893-8e38-ac34600fcd4d","Type":"ContainerStarted","Data":"828cb04e1231792875246ea60327be2d6747449ecc5b9bd5d5c2e79bbd1e5ad0"} Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.453006 4800 scope.go:117] "RemoveContainer" containerID="72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d" Jan 22 00:17:55 crc kubenswrapper[4800]: E0122 00:17:55.453526 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d\": container with ID starting with 72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d not found: ID does not exist" containerID="72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.453557 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d"} err="failed to get container status \"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d\": rpc error: code = NotFound desc = could not find container \"72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d\": container with ID starting with 72991d748bf90e6a98b033bfbe4a0d057477ce5c56fcf05225864b2d1d63c41d not found: ID does not exist" Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.480230 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:55 crc kubenswrapper[4800]: I0122 00:17:55.490567 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 22 00:17:55 crc kubenswrapper[4800]: E0122 00:17:55.495457 4800 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=1939736701449323577, SKID=, AKID=7D:0C:CE:63:0E:3B:61:8B:08:EA:01:3E:D3:A8:86:0F:A8:4F:E5:54 failed: x509: certificate signed by unknown authority" Jan 22 00:17:56 crc kubenswrapper[4800]: I0122 00:17:56.531714 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:17:56 crc kubenswrapper[4800]: I0122 00:17:56.825330 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" path="/var/lib/kubelet/pods/3b5414b4-518d-4c6b-a9cc-f9807a6fe609/volumes" Jan 22 00:17:57 crc kubenswrapper[4800]: I0122 00:17:57.447467 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-2-build" podUID="e63289fb-6cba-4893-8e38-ac34600fcd4d" containerName="git-clone" containerID="cri-o://828cb04e1231792875246ea60327be2d6747449ecc5b9bd5d5c2e79bbd1e5ad0" gracePeriod=30 Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.471081 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_e63289fb-6cba-4893-8e38-ac34600fcd4d/git-clone/0.log" Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.471371 4800 generic.go:334] "Generic (PLEG): container finished" podID="e63289fb-6cba-4893-8e38-ac34600fcd4d" containerID="828cb04e1231792875246ea60327be2d6747449ecc5b9bd5d5c2e79bbd1e5ad0" exitCode=1 Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.471470 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e63289fb-6cba-4893-8e38-ac34600fcd4d","Type":"ContainerDied","Data":"828cb04e1231792875246ea60327be2d6747449ecc5b9bd5d5c2e79bbd1e5ad0"} Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.473026 4800 generic.go:334] "Generic (PLEG): container finished" podID="9791f1db-7616-4146-8651-2dc03b95fdc7" containerID="d45606e2750122f5757f1db7e3dda6bedb703c4e76f4c54534ab86cff6a991bd" exitCode=0 Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.473068 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9791f1db-7616-4146-8651-2dc03b95fdc7","Type":"ContainerDied","Data":"d45606e2750122f5757f1db7e3dda6bedb703c4e76f4c54534ab86cff6a991bd"} Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.850522 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_e63289fb-6cba-4893-8e38-ac34600fcd4d/git-clone/0.log" Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.850832 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999055 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999124 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999153 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999185 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999217 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999239 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999350 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999337 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999412 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999472 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999507 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999557 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6n8r\" (UniqueName: \"kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999597 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:01 crc kubenswrapper[4800]: I0122 00:18:01.999656 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root\") pod \"e63289fb-6cba-4893-8e38-ac34600fcd4d\" (UID: \"e63289fb-6cba-4893-8e38-ac34600fcd4d\") " Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.000424 4800 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.000451 4800 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e63289fb-6cba-4893-8e38-ac34600fcd4d-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.001349 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.001481 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.004161 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.004362 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.004463 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.006334 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.006375 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.018404 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r" (OuterVolumeSpecName: "kube-api-access-l6n8r") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "kube-api-access-l6n8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.018432 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-pull") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "builder-dockercfg-t9bnc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.018394 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-push") pod "e63289fb-6cba-4893-8e38-ac34600fcd4d" (UID: "e63289fb-6cba-4893-8e38-ac34600fcd4d"). InnerVolumeSpecName "builder-dockercfg-t9bnc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103053 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6n8r\" (UniqueName: \"kubernetes.io/projected/e63289fb-6cba-4893-8e38-ac34600fcd4d-kube-api-access-l6n8r\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103089 4800 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103101 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103114 4800 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103124 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-push\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103135 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103145 4800 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103155 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/e63289fb-6cba-4893-8e38-ac34600fcd4d-builder-dockercfg-t9bnc-pull\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103164 4800 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.103173 4800 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e63289fb-6cba-4893-8e38-ac34600fcd4d-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.481797 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_e63289fb-6cba-4893-8e38-ac34600fcd4d/git-clone/0.log" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.481910 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e63289fb-6cba-4893-8e38-ac34600fcd4d","Type":"ContainerDied","Data":"066d77bede158a7763abe903bfbe7c5493be8434015592eb147aa3b362b8948b"} Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.482051 4800 scope.go:117] "RemoveContainer" containerID="828cb04e1231792875246ea60327be2d6747449ecc5b9bd5d5c2e79bbd1e5ad0" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.482050 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.514806 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.522903 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 22 00:18:02 crc kubenswrapper[4800]: I0122 00:18:02.825681 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e63289fb-6cba-4893-8e38-ac34600fcd4d" path="/var/lib/kubelet/pods/e63289fb-6cba-4893-8e38-ac34600fcd4d/volumes" Jan 22 00:18:03 crc kubenswrapper[4800]: I0122 00:18:03.496727 4800 generic.go:334] "Generic (PLEG): container finished" podID="9791f1db-7616-4146-8651-2dc03b95fdc7" containerID="44ced8614065a750a6a087bf7488d49746ea37415152f55738f928349d3c70a1" exitCode=0 Jan 22 00:18:03 crc kubenswrapper[4800]: I0122 00:18:03.496996 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9791f1db-7616-4146-8651-2dc03b95fdc7","Type":"ContainerDied","Data":"44ced8614065a750a6a087bf7488d49746ea37415152f55738f928349d3c70a1"} Jan 22 00:18:04 crc kubenswrapper[4800]: I0122 00:18:04.524389 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9791f1db-7616-4146-8651-2dc03b95fdc7","Type":"ContainerStarted","Data":"18a1510bc15584fdca518078ac2004b3d157ec78913ab849de0567707d79d14b"} Jan 22 00:18:04 crc kubenswrapper[4800]: I0122 00:18:04.524894 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:18:04 crc kubenswrapper[4800]: I0122 00:18:04.553210 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=13.74364143 podStartE2EDuration="32.553186598s" podCreationTimestamp="2026-01-22 00:17:32 +0000 UTC" firstStartedPulling="2026-01-22 00:17:35.372753723 +0000 UTC m=+630.853056761" lastFinishedPulling="2026-01-22 00:17:54.182298891 +0000 UTC m=+649.662601929" observedRunningTime="2026-01-22 00:18:04.551312278 +0000 UTC m=+660.031615316" watchObservedRunningTime="2026-01-22 00:18:04.553186598 +0000 UTC m=+660.033489636" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.044502 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:08 crc kubenswrapper[4800]: E0122 00:18:08.045122 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" containerName="manage-dockerfile" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.045138 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" containerName="manage-dockerfile" Jan 22 00:18:08 crc kubenswrapper[4800]: E0122 00:18:08.045164 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63289fb-6cba-4893-8e38-ac34600fcd4d" containerName="git-clone" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.045172 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63289fb-6cba-4893-8e38-ac34600fcd4d" containerName="git-clone" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.045295 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63289fb-6cba-4893-8e38-ac34600fcd4d" containerName="git-clone" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.045584 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5414b4-518d-4c6b-a9cc-f9807a6fe609" containerName="manage-dockerfile" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.046641 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.052075 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-t9bnc" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.052349 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-ca" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.052385 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-global-ca" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.053914 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-sys-config" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.071953 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093522 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093585 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093612 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093637 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093665 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093691 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093714 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093740 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093775 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093801 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093822 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.093848 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ksmk\" (UniqueName: \"kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194708 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194760 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194787 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194807 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194825 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194845 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194863 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194878 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194918 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194975 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.194997 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.195026 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ksmk\" (UniqueName: \"kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.195198 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.195266 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.195483 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.195962 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.196211 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.196532 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.196650 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.196854 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.197181 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.201504 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.203705 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.218569 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ksmk\" (UniqueName: \"kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk\") pod \"service-telemetry-operator-3-build\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.363016 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.550271 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" event={"ID":"15b2310b-fa78-47b6-a32b-2c3c3984554d","Type":"ContainerStarted","Data":"95804e3ec769ebb08efdadd1ba020ddfae803800f15a0fdaf14e7005818b12ad"} Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.580957 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-rdnvw" podStartSLOduration=8.447042998 podStartE2EDuration="21.580936284s" podCreationTimestamp="2026-01-22 00:17:47 +0000 UTC" firstStartedPulling="2026-01-22 00:17:54.334559994 +0000 UTC m=+649.814863032" lastFinishedPulling="2026-01-22 00:18:07.46845328 +0000 UTC m=+662.948756318" observedRunningTime="2026-01-22 00:18:08.575662933 +0000 UTC m=+664.055965991" watchObservedRunningTime="2026-01-22 00:18:08.580936284 +0000 UTC m=+664.061239322" Jan 22 00:18:08 crc kubenswrapper[4800]: I0122 00:18:08.838343 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:09 crc kubenswrapper[4800]: I0122 00:18:09.559798 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"4eef59a9-35cc-49bd-b72d-da33793356ab","Type":"ContainerStarted","Data":"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4"} Jan 22 00:18:09 crc kubenswrapper[4800]: I0122 00:18:09.560360 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"4eef59a9-35cc-49bd-b72d-da33793356ab","Type":"ContainerStarted","Data":"75eab54739d56c81ec0e7a5316c981f08788b56df2cdc0d7f5d66d4b3882aa64"} Jan 22 00:18:09 crc kubenswrapper[4800]: E0122 00:18:09.631450 4800 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=1939736701449323577, SKID=, AKID=7D:0C:CE:63:0E:3B:61:8B:08:EA:01:3E:D3:A8:86:0F:A8:4F:E5:54 failed: x509: certificate signed by unknown authority" Jan 22 00:18:10 crc kubenswrapper[4800]: I0122 00:18:10.663140 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.569408 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-3-build" podUID="4eef59a9-35cc-49bd-b72d-da33793356ab" containerName="git-clone" containerID="cri-o://4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4" gracePeriod=30 Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.947667 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cpm4r"] Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.948521 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.959732 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.959939 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 22 00:18:11 crc kubenswrapper[4800]: I0122 00:18:11.960064 4800 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7mfdq" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.053630 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9zzc\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-kube-api-access-n9zzc\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.053684 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.115124 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-66zr6"] Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.116113 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.128825 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cpm4r"] Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.129851 4800 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-lmppj" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.138740 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-66zr6"] Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.160974 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fsjr\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-kube-api-access-8fsjr\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.161300 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9zzc\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-kube-api-access-n9zzc\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.161326 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.161343 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.188483 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.188679 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9zzc\" (UniqueName: \"kubernetes.io/projected/55931e16-6b81-4807-aecb-26ae4b773a2a-kube-api-access-n9zzc\") pod \"cert-manager-webhook-f4fb5df64-cpm4r\" (UID: \"55931e16-6b81-4807-aecb-26ae4b773a2a\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.226204 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_4eef59a9-35cc-49bd-b72d-da33793356ab/git-clone/0.log" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.226309 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262252 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262310 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262351 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262379 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262419 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ksmk\" (UniqueName: \"kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262442 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262477 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262502 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262520 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262537 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262596 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262638 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs\") pod \"4eef59a9-35cc-49bd-b72d-da33793356ab\" (UID: \"4eef59a9-35cc-49bd-b72d-da33793356ab\") " Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262751 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262861 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fsjr\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-kube-api-access-8fsjr\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262859 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262953 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.262975 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263012 4800 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263024 4800 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263060 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263350 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263425 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263428 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263451 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.263570 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.269417 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk" (OuterVolumeSpecName: "kube-api-access-8ksmk") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "kube-api-access-8ksmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.269571 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-pull") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "builder-dockercfg-t9bnc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.284924 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-push") pod "4eef59a9-35cc-49bd-b72d-da33793356ab" (UID: "4eef59a9-35cc-49bd-b72d-da33793356ab"). InnerVolumeSpecName "builder-dockercfg-t9bnc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.291228 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.294180 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.315649 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fsjr\" (UniqueName: \"kubernetes.io/projected/acf43b31-0141-4647-a1f8-73e4a1d79917-kube-api-access-8fsjr\") pod \"cert-manager-cainjector-855d9ccff4-66zr6\" (UID: \"acf43b31-0141-4647-a1f8-73e4a1d79917\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364572 4800 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364612 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364628 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/4eef59a9-35cc-49bd-b72d-da33793356ab-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364640 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-push\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364652 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/4eef59a9-35cc-49bd-b72d-da33793356ab-builder-dockercfg-t9bnc-pull\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364662 4800 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364678 4800 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4eef59a9-35cc-49bd-b72d-da33793356ab-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364690 4800 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364699 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ksmk\" (UniqueName: \"kubernetes.io/projected/4eef59a9-35cc-49bd-b72d-da33793356ab-kube-api-access-8ksmk\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.364709 4800 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4eef59a9-35cc-49bd-b72d-da33793356ab-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.446155 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.600585 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_4eef59a9-35cc-49bd-b72d-da33793356ab/git-clone/0.log" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.600994 4800 generic.go:334] "Generic (PLEG): container finished" podID="4eef59a9-35cc-49bd-b72d-da33793356ab" containerID="4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4" exitCode=1 Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.601047 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"4eef59a9-35cc-49bd-b72d-da33793356ab","Type":"ContainerDied","Data":"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4"} Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.601083 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"4eef59a9-35cc-49bd-b72d-da33793356ab","Type":"ContainerDied","Data":"75eab54739d56c81ec0e7a5316c981f08788b56df2cdc0d7f5d66d4b3882aa64"} Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.601106 4800 scope.go:117] "RemoveContainer" containerID="4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.601546 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.634525 4800 scope.go:117] "RemoveContainer" containerID="4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4" Jan 22 00:18:12 crc kubenswrapper[4800]: E0122 00:18:12.636304 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4\": container with ID starting with 4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4 not found: ID does not exist" containerID="4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.636352 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4"} err="failed to get container status \"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4\": rpc error: code = NotFound desc = could not find container \"4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4\": container with ID starting with 4eaa959a20798e22d647bd836906e33a969326eaccff61305993cc2bc9b622f4 not found: ID does not exist" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.655778 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.663741 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.735271 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cpm4r"] Jan 22 00:18:12 crc kubenswrapper[4800]: W0122 00:18:12.745707 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55931e16_6b81_4807_aecb_26ae4b773a2a.slice/crio-bf070b7361f79a19972d54b0528aba44c3abd47c8518ec313fffc6f3fab49dd8 WatchSource:0}: Error finding container bf070b7361f79a19972d54b0528aba44c3abd47c8518ec313fffc6f3fab49dd8: Status 404 returned error can't find the container with id bf070b7361f79a19972d54b0528aba44c3abd47c8518ec313fffc6f3fab49dd8 Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.832316 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eef59a9-35cc-49bd-b72d-da33793356ab" path="/var/lib/kubelet/pods/4eef59a9-35cc-49bd-b72d-da33793356ab/volumes" Jan 22 00:18:12 crc kubenswrapper[4800]: I0122 00:18:12.893399 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-66zr6"] Jan 22 00:18:13 crc kubenswrapper[4800]: I0122 00:18:13.394059 4800 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="9791f1db-7616-4146-8651-2dc03b95fdc7" containerName="elasticsearch" probeResult="failure" output=< Jan 22 00:18:13 crc kubenswrapper[4800]: {"timestamp": "2026-01-22T00:18:13+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 22 00:18:13 crc kubenswrapper[4800]: > Jan 22 00:18:13 crc kubenswrapper[4800]: I0122 00:18:13.611605 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" event={"ID":"acf43b31-0141-4647-a1f8-73e4a1d79917","Type":"ContainerStarted","Data":"0891245b338905a1b0f63a794095597276aa3e19632c654ea38d302512b4e8a7"} Jan 22 00:18:13 crc kubenswrapper[4800]: I0122 00:18:13.614782 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" event={"ID":"55931e16-6b81-4807-aecb-26ae4b773a2a","Type":"ContainerStarted","Data":"bf070b7361f79a19972d54b0528aba44c3abd47c8518ec313fffc6f3fab49dd8"} Jan 22 00:18:19 crc kubenswrapper[4800]: I0122 00:18:19.076172 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.205532 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:22 crc kubenswrapper[4800]: E0122 00:18:22.206240 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eef59a9-35cc-49bd-b72d-da33793356ab" containerName="git-clone" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.206257 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eef59a9-35cc-49bd-b72d-da33793356ab" containerName="git-clone" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.206471 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eef59a9-35cc-49bd-b72d-da33793356ab" containerName="git-clone" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.207574 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.210138 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-t9bnc" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.210205 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-global-ca" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.210415 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-sys-config" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.211069 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-ca" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.232101 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313556 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313607 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313627 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313650 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313715 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313761 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313779 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313798 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313866 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.313968 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8g6\" (UniqueName: \"kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.314016 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.314040 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.414873 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.414979 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415006 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415029 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415056 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415090 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415105 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415121 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415142 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415165 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415190 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8g6\" (UniqueName: \"kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415215 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415689 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.415747 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.416117 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.417523 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.417903 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.418063 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.419904 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.422637 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.427295 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.430393 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8g6\" (UniqueName: \"kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.438640 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.440409 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-4-build\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:22 crc kubenswrapper[4800]: I0122 00:18:22.525535 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:25 crc kubenswrapper[4800]: I0122 00:18:25.689563 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:25 crc kubenswrapper[4800]: W0122 00:18:25.700376 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c7aaa3d_3cbc_4d30_bcfb_f81e46023e3f.slice/crio-4726ed70eb56a84ba776dedfcf4689c1e83d6ddcf8f559c6a73d0ed350bffe77 WatchSource:0}: Error finding container 4726ed70eb56a84ba776dedfcf4689c1e83d6ddcf8f559c6a73d0ed350bffe77: Status 404 returned error can't find the container with id 4726ed70eb56a84ba776dedfcf4689c1e83d6ddcf8f559c6a73d0ed350bffe77 Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.713384 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f","Type":"ContainerStarted","Data":"596ace1f4a37a724d12db631b664a83b0cb56a3f7de82f23b4ff5abc343f3fbf"} Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.713444 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f","Type":"ContainerStarted","Data":"4726ed70eb56a84ba776dedfcf4689c1e83d6ddcf8f559c6a73d0ed350bffe77"} Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.714889 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" event={"ID":"55931e16-6b81-4807-aecb-26ae4b773a2a","Type":"ContainerStarted","Data":"e33106cf6a1daefa1ce1dd8efe3796f69244ec98e74892369f02d77d36d48306"} Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.715025 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.716271 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" event={"ID":"acf43b31-0141-4647-a1f8-73e4a1d79917","Type":"ContainerStarted","Data":"c465b8febadf30a59ec5eafc210d096088dfa1c2f74b2f880287c3b01a899bc8"} Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.771548 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-66zr6" podStartSLOduration=2.384522809 podStartE2EDuration="14.771524147s" podCreationTimestamp="2026-01-22 00:18:12 +0000 UTC" firstStartedPulling="2026-01-22 00:18:12.916076749 +0000 UTC m=+668.396379787" lastFinishedPulling="2026-01-22 00:18:25.303078087 +0000 UTC m=+680.783381125" observedRunningTime="2026-01-22 00:18:26.762341571 +0000 UTC m=+682.242644609" watchObservedRunningTime="2026-01-22 00:18:26.771524147 +0000 UTC m=+682.251827175" Jan 22 00:18:26 crc kubenswrapper[4800]: I0122 00:18:26.780775 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" podStartSLOduration=3.214540049 podStartE2EDuration="15.780746863s" podCreationTimestamp="2026-01-22 00:18:11 +0000 UTC" firstStartedPulling="2026-01-22 00:18:12.751473852 +0000 UTC m=+668.231776890" lastFinishedPulling="2026-01-22 00:18:25.317680666 +0000 UTC m=+680.797983704" observedRunningTime="2026-01-22 00:18:26.777167688 +0000 UTC m=+682.257470726" watchObservedRunningTime="2026-01-22 00:18:26.780746863 +0000 UTC m=+682.261049901" Jan 22 00:18:26 crc kubenswrapper[4800]: E0122 00:18:26.790872 4800 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=1939736701449323577, SKID=, AKID=7D:0C:CE:63:0E:3B:61:8B:08:EA:01:3E:D3:A8:86:0F:A8:4F:E5:54 failed: x509: certificate signed by unknown authority" Jan 22 00:18:27 crc kubenswrapper[4800]: I0122 00:18:27.835663 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:28 crc kubenswrapper[4800]: I0122 00:18:28.731880 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-4-build" podUID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" containerName="git-clone" containerID="cri-o://596ace1f4a37a724d12db631b664a83b0cb56a3f7de82f23b4ff5abc343f3fbf" gracePeriod=30 Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.747787 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f/git-clone/0.log" Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.748089 4800 generic.go:334] "Generic (PLEG): container finished" podID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" containerID="596ace1f4a37a724d12db631b664a83b0cb56a3f7de82f23b4ff5abc343f3fbf" exitCode=1 Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.748122 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f","Type":"ContainerDied","Data":"596ace1f4a37a724d12db631b664a83b0cb56a3f7de82f23b4ff5abc343f3fbf"} Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.851565 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mss24"] Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.853719 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.869351 4800 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7zrgl" Jan 22 00:18:30 crc kubenswrapper[4800]: I0122 00:18:30.884015 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mss24"] Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.045974 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbc5r\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-kube-api-access-hbc5r\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.046081 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-bound-sa-token\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.147032 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-bound-sa-token\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.147539 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbc5r\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-kube-api-access-hbc5r\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.196735 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-bound-sa-token\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.197097 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbc5r\" (UniqueName: \"kubernetes.io/projected/606d52f5-7d6d-4d7c-a546-c1ceef55d61a-kube-api-access-hbc5r\") pod \"cert-manager-86cb77c54b-mss24\" (UID: \"606d52f5-7d6d-4d7c-a546-c1ceef55d61a\") " pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.398109 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f/git-clone/0.log" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.398198 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.496260 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-mss24" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.553476 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.553961 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.553995 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554011 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c8g6\" (UniqueName: \"kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554033 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554052 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554067 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554102 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554131 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554161 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554203 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554233 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push\") pod \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\" (UID: \"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f\") " Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554547 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554670 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.554808 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555016 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555020 4800 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555071 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555107 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555134 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555412 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555500 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.555669 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.561089 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-pull") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "builder-dockercfg-t9bnc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.561349 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6" (OuterVolumeSpecName: "kube-api-access-8c8g6") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "kube-api-access-8c8g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.562571 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-push") pod "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" (UID: "0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f"). InnerVolumeSpecName "builder-dockercfg-t9bnc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656421 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656473 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-push\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656484 4800 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656494 4800 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656505 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c8g6\" (UniqueName: \"kubernetes.io/projected/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-kube-api-access-8c8g6\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656514 4800 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656523 4800 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656533 4800 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656542 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-builder-dockercfg-t9bnc-pull\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.656551 4800 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.759505 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f/git-clone/0.log" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.759570 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f","Type":"ContainerDied","Data":"4726ed70eb56a84ba776dedfcf4689c1e83d6ddcf8f559c6a73d0ed350bffe77"} Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.759624 4800 scope.go:117] "RemoveContainer" containerID="596ace1f4a37a724d12db631b664a83b0cb56a3f7de82f23b4ff5abc343f3fbf" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.759804 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.795056 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.807251 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Jan 22 00:18:31 crc kubenswrapper[4800]: I0122 00:18:31.953418 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-mss24"] Jan 22 00:18:32 crc kubenswrapper[4800]: I0122 00:18:32.297697 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-cpm4r" Jan 22 00:18:32 crc kubenswrapper[4800]: I0122 00:18:32.774204 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-mss24" event={"ID":"606d52f5-7d6d-4d7c-a546-c1ceef55d61a","Type":"ContainerStarted","Data":"41b4ed1c3b6c31e880c27c0964b5e0d7c49f44adec130d98452338aa533507bd"} Jan 22 00:18:32 crc kubenswrapper[4800]: I0122 00:18:32.774265 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-mss24" event={"ID":"606d52f5-7d6d-4d7c-a546-c1ceef55d61a","Type":"ContainerStarted","Data":"ddcc83b8557f69e9d22049d052b715059c0377cf6ccbf1482cd7a3b14b39c384"} Jan 22 00:18:32 crc kubenswrapper[4800]: I0122 00:18:32.799367 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-mss24" podStartSLOduration=2.799341351 podStartE2EDuration="2.799341351s" podCreationTimestamp="2026-01-22 00:18:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:18:32.796164046 +0000 UTC m=+688.276467104" watchObservedRunningTime="2026-01-22 00:18:32.799341351 +0000 UTC m=+688.279644389" Jan 22 00:18:32 crc kubenswrapper[4800]: I0122 00:18:32.825665 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" path="/var/lib/kubelet/pods/0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f/volumes" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.335513 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:39 crc kubenswrapper[4800]: E0122 00:18:39.336440 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" containerName="git-clone" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.336456 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" containerName="git-clone" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.336623 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7aaa3d-3cbc-4d30-bcfb-f81e46023e3f" containerName="git-clone" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.337694 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.340341 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-global-ca" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.340394 4800 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-t9bnc" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.340879 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-ca" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.352200 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.354669 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-sys-config" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.465909 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.465955 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.465998 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466018 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466036 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466055 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466077 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466095 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466112 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466135 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws6mc\" (UniqueName: \"kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466170 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.466184 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567527 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567581 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567661 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567687 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567707 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567739 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws6mc\" (UniqueName: \"kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567786 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567809 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567837 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567859 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567919 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.567944 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568220 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568228 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568287 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568362 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568626 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568861 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.568924 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.569047 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.569326 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.573664 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.574094 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.587250 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws6mc\" (UniqueName: \"kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc\") pod \"service-telemetry-operator-5-build\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.659120 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:39 crc kubenswrapper[4800]: I0122 00:18:39.899301 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:40 crc kubenswrapper[4800]: I0122 00:18:40.831950 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"5427c52a-b97d-4cdc-9a78-a4356bddf757","Type":"ContainerStarted","Data":"88009f14ca8412ef081bf48d9277469e7657e06652b3057efa66be67f1aa6259"} Jan 22 00:18:41 crc kubenswrapper[4800]: I0122 00:18:41.836072 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"5427c52a-b97d-4cdc-9a78-a4356bddf757","Type":"ContainerStarted","Data":"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8"} Jan 22 00:18:41 crc kubenswrapper[4800]: E0122 00:18:41.893361 4800 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=1939736701449323577, SKID=, AKID=7D:0C:CE:63:0E:3B:61:8B:08:EA:01:3E:D3:A8:86:0F:A8:4F:E5:54 failed: x509: certificate signed by unknown authority" Jan 22 00:18:42 crc kubenswrapper[4800]: I0122 00:18:42.919939 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:43 crc kubenswrapper[4800]: I0122 00:18:43.846198 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-5-build" podUID="5427c52a-b97d-4cdc-9a78-a4356bddf757" containerName="git-clone" containerID="cri-o://8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8" gracePeriod=30 Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.191529 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_5427c52a-b97d-4cdc-9a78-a4356bddf757/git-clone/0.log" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.192034 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350513 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350588 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350639 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350660 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350680 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350716 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350732 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350749 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws6mc\" (UniqueName: \"kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350800 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350819 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350848 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.350872 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles\") pod \"5427c52a-b97d-4cdc-9a78-a4356bddf757\" (UID: \"5427c52a-b97d-4cdc-9a78-a4356bddf757\") " Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.351268 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.351742 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352003 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352056 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352181 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352216 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352399 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.352808 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.353301 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.357674 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc" (OuterVolumeSpecName: "kube-api-access-ws6mc") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "kube-api-access-ws6mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.357706 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-pull") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "builder-dockercfg-t9bnc-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.360447 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push" (OuterVolumeSpecName: "builder-dockercfg-t9bnc-push") pod "5427c52a-b97d-4cdc-9a78-a4356bddf757" (UID: "5427c52a-b97d-4cdc-9a78-a4356bddf757"). InnerVolumeSpecName "builder-dockercfg-t9bnc-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451871 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451920 4800 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451933 4800 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451947 4800 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451958 4800 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451970 4800 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451981 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-push\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-push\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.451993 4800 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.452002 4800 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5427c52a-b97d-4cdc-9a78-a4356bddf757-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.452010 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws6mc\" (UniqueName: \"kubernetes.io/projected/5427c52a-b97d-4cdc-9a78-a4356bddf757-kube-api-access-ws6mc\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.452018 4800 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-t9bnc-pull\" (UniqueName: \"kubernetes.io/secret/5427c52a-b97d-4cdc-9a78-a4356bddf757-builder-dockercfg-t9bnc-pull\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.452026 4800 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5427c52a-b97d-4cdc-9a78-a4356bddf757-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.852717 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_5427c52a-b97d-4cdc-9a78-a4356bddf757/git-clone/0.log" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.852767 4800 generic.go:334] "Generic (PLEG): container finished" podID="5427c52a-b97d-4cdc-9a78-a4356bddf757" containerID="8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8" exitCode=1 Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.852799 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"5427c52a-b97d-4cdc-9a78-a4356bddf757","Type":"ContainerDied","Data":"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8"} Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.852830 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"5427c52a-b97d-4cdc-9a78-a4356bddf757","Type":"ContainerDied","Data":"88009f14ca8412ef081bf48d9277469e7657e06652b3057efa66be67f1aa6259"} Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.852853 4800 scope.go:117] "RemoveContainer" containerID="8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.853010 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.876434 4800 scope.go:117] "RemoveContainer" containerID="8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.877222 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:44 crc kubenswrapper[4800]: E0122 00:18:44.877609 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8\": container with ID starting with 8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8 not found: ID does not exist" containerID="8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.878614 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8"} err="failed to get container status \"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8\": rpc error: code = NotFound desc = could not find container \"8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8\": container with ID starting with 8d00dba25b70f931b57b8faf7e0b26d7712e8c5342f366f1f5426730cfbfe9f8 not found: ID does not exist" Jan 22 00:18:44 crc kubenswrapper[4800]: I0122 00:18:44.884239 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Jan 22 00:18:46 crc kubenswrapper[4800]: I0122 00:18:46.827531 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5427c52a-b97d-4cdc-9a78-a4356bddf757" path="/var/lib/kubelet/pods/5427c52a-b97d-4cdc-9a78-a4356bddf757/volumes" Jan 22 00:18:49 crc kubenswrapper[4800]: I0122 00:18:49.452382 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:18:49 crc kubenswrapper[4800]: I0122 00:18:49.452757 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:19:19 crc kubenswrapper[4800]: I0122 00:19:19.451503 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:19:19 crc kubenswrapper[4800]: I0122 00:19:19.452123 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.645939 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wrmd4/must-gather-wcv62"] Jan 22 00:19:31 crc kubenswrapper[4800]: E0122 00:19:31.646879 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5427c52a-b97d-4cdc-9a78-a4356bddf757" containerName="git-clone" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.646901 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="5427c52a-b97d-4cdc-9a78-a4356bddf757" containerName="git-clone" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.647057 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="5427c52a-b97d-4cdc-9a78-a4356bddf757" containerName="git-clone" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.647907 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.655230 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wrmd4"/"openshift-service-ca.crt" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.655531 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wrmd4"/"kube-root-ca.crt" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.656048 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wrmd4"/"default-dockercfg-wnjvd" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.665333 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wrmd4/must-gather-wcv62"] Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.768948 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntx59\" (UniqueName: \"kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.769037 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.869928 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntx59\" (UniqueName: \"kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.870339 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.870751 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.898545 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntx59\" (UniqueName: \"kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59\") pod \"must-gather-wcv62\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:31 crc kubenswrapper[4800]: I0122 00:19:31.971296 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:19:32 crc kubenswrapper[4800]: I0122 00:19:32.186205 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wrmd4/must-gather-wcv62"] Jan 22 00:19:33 crc kubenswrapper[4800]: I0122 00:19:33.173617 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wrmd4/must-gather-wcv62" event={"ID":"51656a67-bab8-48ce-bcd1-89209dfa34a0","Type":"ContainerStarted","Data":"18b286d8aba56d44c36796a34b05626606aebf7536144b95a18f685973cd5f8f"} Jan 22 00:19:40 crc kubenswrapper[4800]: I0122 00:19:40.262506 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wrmd4/must-gather-wcv62" event={"ID":"51656a67-bab8-48ce-bcd1-89209dfa34a0","Type":"ContainerStarted","Data":"3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6"} Jan 22 00:19:40 crc kubenswrapper[4800]: I0122 00:19:40.263041 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wrmd4/must-gather-wcv62" event={"ID":"51656a67-bab8-48ce-bcd1-89209dfa34a0","Type":"ContainerStarted","Data":"e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9"} Jan 22 00:19:41 crc kubenswrapper[4800]: E0122 00:19:41.640725 4800 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=1939736701449323577, SKID=, AKID=7D:0C:CE:63:0E:3B:61:8B:08:EA:01:3E:D3:A8:86:0F:A8:4F:E5:54 failed: x509: certificate signed by unknown authority" Jan 22 00:19:43 crc kubenswrapper[4800]: I0122 00:19:43.956270 4800 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 00:19:49 crc kubenswrapper[4800]: I0122 00:19:49.451829 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:19:49 crc kubenswrapper[4800]: I0122 00:19:49.452425 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:19:49 crc kubenswrapper[4800]: I0122 00:19:49.452477 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:19:49 crc kubenswrapper[4800]: I0122 00:19:49.453227 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:19:49 crc kubenswrapper[4800]: I0122 00:19:49.453283 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c" gracePeriod=600 Jan 22 00:19:51 crc kubenswrapper[4800]: I0122 00:19:51.330496 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c" exitCode=0 Jan 22 00:19:51 crc kubenswrapper[4800]: I0122 00:19:51.330593 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c"} Jan 22 00:19:51 crc kubenswrapper[4800]: I0122 00:19:51.331209 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea"} Jan 22 00:19:51 crc kubenswrapper[4800]: I0122 00:19:51.331238 4800 scope.go:117] "RemoveContainer" containerID="a5ccb73ed4cc054d2e414fbb0fd8f15044e558182ae1692ae9c6ce9a186c0861" Jan 22 00:19:51 crc kubenswrapper[4800]: I0122 00:19:51.363009 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wrmd4/must-gather-wcv62" podStartSLOduration=12.908637388 podStartE2EDuration="20.362979105s" podCreationTimestamp="2026-01-22 00:19:31 +0000 UTC" firstStartedPulling="2026-01-22 00:19:32.19886971 +0000 UTC m=+747.679172748" lastFinishedPulling="2026-01-22 00:19:39.653211427 +0000 UTC m=+755.133514465" observedRunningTime="2026-01-22 00:19:40.279268237 +0000 UTC m=+755.759571275" watchObservedRunningTime="2026-01-22 00:19:51.362979105 +0000 UTC m=+766.843282143" Jan 22 00:19:52 crc kubenswrapper[4800]: I0122 00:19:52.470900 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6grhd_cf201413-6366-4074-8cc2-90819f77229c/control-plane-machine-set-operator/0.log" Jan 22 00:19:52 crc kubenswrapper[4800]: I0122 00:19:52.486828 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hbblt_72de5335-31e7-4e8a-8dee-2b5c8afdeab9/kube-rbac-proxy/0.log" Jan 22 00:19:52 crc kubenswrapper[4800]: I0122 00:19:52.495752 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hbblt_72de5335-31e7-4e8a-8dee-2b5c8afdeab9/machine-api-operator/0.log" Jan 22 00:19:57 crc kubenswrapper[4800]: I0122 00:19:57.566418 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mss24_606d52f5-7d6d-4d7c-a546-c1ceef55d61a/cert-manager-controller/0.log" Jan 22 00:19:57 crc kubenswrapper[4800]: I0122 00:19:57.579878 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-66zr6_acf43b31-0141-4647-a1f8-73e4a1d79917/cert-manager-cainjector/0.log" Jan 22 00:19:57 crc kubenswrapper[4800]: I0122 00:19:57.594921 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cpm4r_55931e16-6b81-4807-aecb-26ae4b773a2a/cert-manager-webhook/0.log" Jan 22 00:20:02 crc kubenswrapper[4800]: I0122 00:20:02.689037 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-wqvbx_68953295-7f82-41aa-a591-a572d99b9210/prometheus-operator/0.log" Jan 22 00:20:02 crc kubenswrapper[4800]: I0122 00:20:02.703777 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq_fb96f099-5861-45a5-89e7-2ab5c80a90cb/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:02 crc kubenswrapper[4800]: I0122 00:20:02.715969 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x_0771650e-3769-4dd8-baff-e6fc99f841a4/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:02 crc kubenswrapper[4800]: I0122 00:20:02.732830 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-zhrc2_d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3/operator/0.log" Jan 22 00:20:02 crc kubenswrapper[4800]: I0122 00:20:02.745541 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-p9zm8_2b7c78d4-97fb-48c5-931d-8477f4876451/perses-operator/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.859447 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk_48aa1c61-a36e-4c82-8812-37ceb4f45d59/extract/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.873598 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk_48aa1c61-a36e-4c82-8812-37ceb4f45d59/util/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.902272 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931al98vk_48aa1c61-a36e-4c82-8812-37ceb4f45d59/pull/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.911168 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x_358633f6-ece5-4055-98a2-25f2805e3e06/extract/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.917306 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x_358633f6-ece5-4055-98a2-25f2805e3e06/util/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.925717 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f95c7x_358633f6-ece5-4055-98a2-25f2805e3e06/pull/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.938013 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg_bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b/extract/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.946248 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg_bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b/util/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.955118 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5egwnsg_bc2feeb1-c6a6-4b6a-956f-e5fd2f72d74b/pull/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.966685 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj_9e04ccf1-c955-4975-acb3-5e92cbab3469/extract/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.973647 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj_9e04ccf1-c955-4975-acb3-5e92cbab3469/util/0.log" Jan 22 00:20:07 crc kubenswrapper[4800]: I0122 00:20:07.982540 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f082wvpj_9e04ccf1-c955-4975-acb3-5e92cbab3469/pull/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.184866 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgpdn_f35183da-e8fc-4399-a3fa-20b466813639/registry-server/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.192110 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgpdn_f35183da-e8fc-4399-a3fa-20b466813639/extract-utilities/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.199772 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgpdn_f35183da-e8fc-4399-a3fa-20b466813639/extract-content/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.419718 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h7qwp_994f0255-587e-4ad7-a84f-c4c8a0e06184/registry-server/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.424588 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h7qwp_994f0255-587e-4ad7-a84f-c4c8a0e06184/extract-utilities/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.432429 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-h7qwp_994f0255-587e-4ad7-a84f-c4c8a0e06184/extract-content/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.460358 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8d98k_a816e793-2455-4446-85dd-6437bc1d6ee7/marketplace-operator/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.656061 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x8jld_ecdd722c-78ea-4e3e-bd54-31e53b1177da/registry-server/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.662081 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x8jld_ecdd722c-78ea-4e3e-bd54-31e53b1177da/extract-utilities/0.log" Jan 22 00:20:08 crc kubenswrapper[4800]: I0122 00:20:08.669136 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x8jld_ecdd722c-78ea-4e3e-bd54-31e53b1177da/extract-content/0.log" Jan 22 00:20:12 crc kubenswrapper[4800]: I0122 00:20:12.877539 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-wqvbx_68953295-7f82-41aa-a591-a572d99b9210/prometheus-operator/0.log" Jan 22 00:20:12 crc kubenswrapper[4800]: I0122 00:20:12.895923 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq_fb96f099-5861-45a5-89e7-2ab5c80a90cb/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:12 crc kubenswrapper[4800]: I0122 00:20:12.906904 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x_0771650e-3769-4dd8-baff-e6fc99f841a4/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:12 crc kubenswrapper[4800]: I0122 00:20:12.942318 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-zhrc2_d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3/operator/0.log" Jan 22 00:20:12 crc kubenswrapper[4800]: I0122 00:20:12.951743 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-p9zm8_2b7c78d4-97fb-48c5-931d-8477f4876451/perses-operator/0.log" Jan 22 00:20:20 crc kubenswrapper[4800]: I0122 00:20:20.999077 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-wqvbx_68953295-7f82-41aa-a591-a572d99b9210/prometheus-operator/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.010291 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-788mq_fb96f099-5861-45a5-89e7-2ab5c80a90cb/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.020835 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b9cbbd855-qs86x_0771650e-3769-4dd8-baff-e6fc99f841a4/prometheus-operator-admission-webhook/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.037899 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-zhrc2_d52996fb-0fe0-4d5d-b9ac-3ae434ddfbc3/operator/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.054668 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-p9zm8_2b7c78d4-97fb-48c5-931d-8477f4876451/perses-operator/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.164827 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mss24_606d52f5-7d6d-4d7c-a546-c1ceef55d61a/cert-manager-controller/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.187915 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-66zr6_acf43b31-0141-4647-a1f8-73e4a1d79917/cert-manager-cainjector/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.198819 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cpm4r_55931e16-6b81-4807-aecb-26ae4b773a2a/cert-manager-webhook/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.650945 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-mss24_606d52f5-7d6d-4d7c-a546-c1ceef55d61a/cert-manager-controller/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.671272 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-66zr6_acf43b31-0141-4647-a1f8-73e4a1d79917/cert-manager-cainjector/0.log" Jan 22 00:20:21 crc kubenswrapper[4800]: I0122 00:20:21.682298 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cpm4r_55931e16-6b81-4807-aecb-26ae4b773a2a/cert-manager-webhook/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.066733 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6grhd_cf201413-6366-4074-8cc2-90819f77229c/control-plane-machine-set-operator/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.080366 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hbblt_72de5335-31e7-4e8a-8dee-2b5c8afdeab9/kube-rbac-proxy/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.090470 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hbblt_72de5335-31e7-4e8a-8dee-2b5c8afdeab9/machine-api-operator/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.626222 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elastic-operator-545b84bc9b-ln8zq_a247bbf3-14dc-4b0a-8ae2-27272b6b6f16/manager/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.648023 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_9791f1db-7616-4146-8651-2dc03b95fdc7/elasticsearch/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.656665 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_9791f1db-7616-4146-8651-2dc03b95fdc7/elastic-internal-init-filesystem/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.660834 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_9791f1db-7616-4146-8651-2dc03b95fdc7/elastic-internal-suspend/0.log" Jan 22 00:20:22 crc kubenswrapper[4800]: I0122 00:20:22.671077 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_interconnect-operator-5bb49f789d-m47bj_e1c2c5e3-e04e-4626-b7b9-bd0b7ee3ded3/interconnect-operator/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.490772 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/kube-multus-additional-cni-plugins/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.499603 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/egress-router-binary-copy/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.507153 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/cni-plugins/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.515586 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/bond-cni-plugin/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.523075 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/routeoverride-cni/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.533924 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/whereabouts-cni-bincopy/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.542375 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-bv9wt_fdb45e62-a601-42e3-988e-4834834ecb6b/whereabouts-cni/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.553641 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-wv42v_dadf4421-75bf-4cb8-8d63-040b446ecf0f/multus-admission-controller/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.561534 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-wv42v_dadf4421-75bf-4cb8-8d63-040b446ecf0f/kube-rbac-proxy/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.597265 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/3.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.621649 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vfgd4_aff2b22c-b87d-47c9-b5a0-6e5772022488/kube-multus/2.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.638706 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-svkb9_ca19c294-fff4-4fe6-a1a4-25e04d0624b3/network-metrics-daemon/0.log" Jan 22 00:20:23 crc kubenswrapper[4800]: I0122 00:20:23.643638 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-svkb9_ca19c294-fff4-4fe6-a1a4-25e04d0624b3/kube-rbac-proxy/0.log" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.496025 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.497928 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.501498 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.594767 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.594861 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwsbx\" (UniqueName: \"kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.594942 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.696311 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.696714 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwsbx\" (UniqueName: \"kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.696956 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.697015 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.697772 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.720069 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwsbx\" (UniqueName: \"kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx\") pod \"community-operators-584rz\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:09 crc kubenswrapper[4800]: I0122 00:22:09.838565 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:10 crc kubenswrapper[4800]: I0122 00:22:10.322795 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:11 crc kubenswrapper[4800]: I0122 00:22:11.236931 4800 generic.go:334] "Generic (PLEG): container finished" podID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerID="afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49" exitCode=0 Jan 22 00:22:11 crc kubenswrapper[4800]: I0122 00:22:11.239135 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerDied","Data":"afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49"} Jan 22 00:22:11 crc kubenswrapper[4800]: I0122 00:22:11.239207 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerStarted","Data":"257de740161e2d1250bc0d84d78b651a0d32f71b0af93e773b6ebfeb1aa17cab"} Jan 22 00:22:11 crc kubenswrapper[4800]: I0122 00:22:11.250202 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:22:12 crc kubenswrapper[4800]: I0122 00:22:12.245303 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerStarted","Data":"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956"} Jan 22 00:22:13 crc kubenswrapper[4800]: I0122 00:22:13.252206 4800 generic.go:334] "Generic (PLEG): container finished" podID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerID="7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956" exitCode=0 Jan 22 00:22:13 crc kubenswrapper[4800]: I0122 00:22:13.252291 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerDied","Data":"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956"} Jan 22 00:22:14 crc kubenswrapper[4800]: I0122 00:22:14.260813 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerStarted","Data":"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba"} Jan 22 00:22:14 crc kubenswrapper[4800]: I0122 00:22:14.284503 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-584rz" podStartSLOduration=2.9089941169999998 podStartE2EDuration="5.28448033s" podCreationTimestamp="2026-01-22 00:22:09 +0000 UTC" firstStartedPulling="2026-01-22 00:22:11.249880912 +0000 UTC m=+906.730183960" lastFinishedPulling="2026-01-22 00:22:13.625367135 +0000 UTC m=+909.105670173" observedRunningTime="2026-01-22 00:22:14.282431735 +0000 UTC m=+909.762734773" watchObservedRunningTime="2026-01-22 00:22:14.28448033 +0000 UTC m=+909.764783368" Jan 22 00:22:19 crc kubenswrapper[4800]: I0122 00:22:19.451986 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:22:19 crc kubenswrapper[4800]: I0122 00:22:19.452600 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:22:19 crc kubenswrapper[4800]: I0122 00:22:19.838771 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:19 crc kubenswrapper[4800]: I0122 00:22:19.838820 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:19 crc kubenswrapper[4800]: I0122 00:22:19.885772 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:20 crc kubenswrapper[4800]: I0122 00:22:20.336291 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:20 crc kubenswrapper[4800]: I0122 00:22:20.375466 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.309535 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-584rz" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="registry-server" containerID="cri-o://d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba" gracePeriod=2 Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.670502 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.788928 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content\") pod \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.806098 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities\") pod \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.806169 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwsbx\" (UniqueName: \"kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx\") pod \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\" (UID: \"512600ad-f974-4c9a-ae2e-5874c3ef43ac\") " Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.807052 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities" (OuterVolumeSpecName: "utilities") pod "512600ad-f974-4c9a-ae2e-5874c3ef43ac" (UID: "512600ad-f974-4c9a-ae2e-5874c3ef43ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.812346 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx" (OuterVolumeSpecName: "kube-api-access-qwsbx") pod "512600ad-f974-4c9a-ae2e-5874c3ef43ac" (UID: "512600ad-f974-4c9a-ae2e-5874c3ef43ac"). InnerVolumeSpecName "kube-api-access-qwsbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.848409 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "512600ad-f974-4c9a-ae2e-5874c3ef43ac" (UID: "512600ad-f974-4c9a-ae2e-5874c3ef43ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.908075 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.908121 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512600ad-f974-4c9a-ae2e-5874c3ef43ac-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:22 crc kubenswrapper[4800]: I0122 00:22:22.908146 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwsbx\" (UniqueName: \"kubernetes.io/projected/512600ad-f974-4c9a-ae2e-5874c3ef43ac-kube-api-access-qwsbx\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.316610 4800 generic.go:334] "Generic (PLEG): container finished" podID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerID="d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba" exitCode=0 Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.316654 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerDied","Data":"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba"} Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.316674 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-584rz" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.316698 4800 scope.go:117] "RemoveContainer" containerID="d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.316684 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-584rz" event={"ID":"512600ad-f974-4c9a-ae2e-5874c3ef43ac","Type":"ContainerDied","Data":"257de740161e2d1250bc0d84d78b651a0d32f71b0af93e773b6ebfeb1aa17cab"} Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.341621 4800 scope.go:117] "RemoveContainer" containerID="7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.360781 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.366585 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-584rz"] Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.374168 4800 scope.go:117] "RemoveContainer" containerID="afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.409785 4800 scope.go:117] "RemoveContainer" containerID="d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba" Jan 22 00:22:23 crc kubenswrapper[4800]: E0122 00:22:23.410324 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba\": container with ID starting with d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba not found: ID does not exist" containerID="d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.410364 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba"} err="failed to get container status \"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba\": rpc error: code = NotFound desc = could not find container \"d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba\": container with ID starting with d372b878753b5b6248ca62297bc194fd7e3c7d4a4a8d937feec9c7b7b0e97aba not found: ID does not exist" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.410395 4800 scope.go:117] "RemoveContainer" containerID="7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956" Jan 22 00:22:23 crc kubenswrapper[4800]: E0122 00:22:23.410610 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956\": container with ID starting with 7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956 not found: ID does not exist" containerID="7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.410639 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956"} err="failed to get container status \"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956\": rpc error: code = NotFound desc = could not find container \"7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956\": container with ID starting with 7a318e15db8c5c4125a99d22f797b1a51818cc689ae5a80f45641cb5c9d6f956 not found: ID does not exist" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.410658 4800 scope.go:117] "RemoveContainer" containerID="afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49" Jan 22 00:22:23 crc kubenswrapper[4800]: E0122 00:22:23.410880 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49\": container with ID starting with afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49 not found: ID does not exist" containerID="afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49" Jan 22 00:22:23 crc kubenswrapper[4800]: I0122 00:22:23.410922 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49"} err="failed to get container status \"afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49\": rpc error: code = NotFound desc = could not find container \"afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49\": container with ID starting with afcbe4600ce21364ff5243e0908cabf39fba7b281f2d3bcc73df17a6e3394e49 not found: ID does not exist" Jan 22 00:22:24 crc kubenswrapper[4800]: I0122 00:22:24.836090 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" path="/var/lib/kubelet/pods/512600ad-f974-4c9a-ae2e-5874c3ef43ac/volumes" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.897254 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:28 crc kubenswrapper[4800]: E0122 00:22:28.898554 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="extract-utilities" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.898657 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="extract-utilities" Jan 22 00:22:28 crc kubenswrapper[4800]: E0122 00:22:28.898733 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="registry-server" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.898818 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="registry-server" Jan 22 00:22:28 crc kubenswrapper[4800]: E0122 00:22:28.898924 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="extract-content" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.899013 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="extract-content" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.899281 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="512600ad-f974-4c9a-ae2e-5874c3ef43ac" containerName="registry-server" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.900317 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:28 crc kubenswrapper[4800]: I0122 00:22:28.907674 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.019872 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5xj\" (UniqueName: \"kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.019971 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.019994 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.121471 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc5xj\" (UniqueName: \"kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.121750 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.121824 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.122336 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.122394 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.140611 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc5xj\" (UniqueName: \"kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj\") pod \"redhat-operators-q48wn\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.223605 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:29 crc kubenswrapper[4800]: I0122 00:22:29.430415 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:30 crc kubenswrapper[4800]: I0122 00:22:30.374824 4800 generic.go:334] "Generic (PLEG): container finished" podID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerID="5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24" exitCode=0 Jan 22 00:22:30 crc kubenswrapper[4800]: I0122 00:22:30.374878 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerDied","Data":"5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24"} Jan 22 00:22:30 crc kubenswrapper[4800]: I0122 00:22:30.375094 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerStarted","Data":"b9d494519d3ed802ecea2929d1da9be26d699070121880f0f119456808e84a78"} Jan 22 00:22:32 crc kubenswrapper[4800]: I0122 00:22:32.389504 4800 generic.go:334] "Generic (PLEG): container finished" podID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerID="6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112" exitCode=0 Jan 22 00:22:32 crc kubenswrapper[4800]: I0122 00:22:32.389578 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerDied","Data":"6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112"} Jan 22 00:22:33 crc kubenswrapper[4800]: I0122 00:22:33.398937 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerStarted","Data":"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1"} Jan 22 00:22:33 crc kubenswrapper[4800]: I0122 00:22:33.425088 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q48wn" podStartSLOduration=2.854322584 podStartE2EDuration="5.425071175s" podCreationTimestamp="2026-01-22 00:22:28 +0000 UTC" firstStartedPulling="2026-01-22 00:22:30.377996353 +0000 UTC m=+925.858299391" lastFinishedPulling="2026-01-22 00:22:32.948744944 +0000 UTC m=+928.429047982" observedRunningTime="2026-01-22 00:22:33.423699669 +0000 UTC m=+928.904002707" watchObservedRunningTime="2026-01-22 00:22:33.425071175 +0000 UTC m=+928.905374213" Jan 22 00:22:39 crc kubenswrapper[4800]: I0122 00:22:39.224683 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:39 crc kubenswrapper[4800]: I0122 00:22:39.225673 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:39 crc kubenswrapper[4800]: I0122 00:22:39.273180 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:39 crc kubenswrapper[4800]: I0122 00:22:39.494733 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:40 crc kubenswrapper[4800]: I0122 00:22:40.522872 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:41 crc kubenswrapper[4800]: I0122 00:22:41.451809 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q48wn" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="registry-server" containerID="cri-o://8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1" gracePeriod=2 Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.341050 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.418525 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc5xj\" (UniqueName: \"kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj\") pod \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.419098 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content\") pod \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.419208 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities\") pod \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\" (UID: \"da7c939b-1cc4-46a7-af60-968d3efbdcd9\") " Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.420146 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities" (OuterVolumeSpecName: "utilities") pod "da7c939b-1cc4-46a7-af60-968d3efbdcd9" (UID: "da7c939b-1cc4-46a7-af60-968d3efbdcd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.426111 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj" (OuterVolumeSpecName: "kube-api-access-nc5xj") pod "da7c939b-1cc4-46a7-af60-968d3efbdcd9" (UID: "da7c939b-1cc4-46a7-af60-968d3efbdcd9"). InnerVolumeSpecName "kube-api-access-nc5xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.460646 4800 generic.go:334] "Generic (PLEG): container finished" podID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerID="8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1" exitCode=0 Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.460709 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerDied","Data":"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1"} Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.460747 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q48wn" event={"ID":"da7c939b-1cc4-46a7-af60-968d3efbdcd9","Type":"ContainerDied","Data":"b9d494519d3ed802ecea2929d1da9be26d699070121880f0f119456808e84a78"} Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.460747 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q48wn" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.460767 4800 scope.go:117] "RemoveContainer" containerID="8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.478552 4800 scope.go:117] "RemoveContainer" containerID="6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.500113 4800 scope.go:117] "RemoveContainer" containerID="5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.518337 4800 scope.go:117] "RemoveContainer" containerID="8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1" Jan 22 00:22:42 crc kubenswrapper[4800]: E0122 00:22:42.518849 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1\": container with ID starting with 8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1 not found: ID does not exist" containerID="8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.518898 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1"} err="failed to get container status \"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1\": rpc error: code = NotFound desc = could not find container \"8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1\": container with ID starting with 8cee52900a06c8fb2285fd8ed7e39c9bdf8955d4d540fe02df963fea329331d1 not found: ID does not exist" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.518924 4800 scope.go:117] "RemoveContainer" containerID="6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112" Jan 22 00:22:42 crc kubenswrapper[4800]: E0122 00:22:42.519281 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112\": container with ID starting with 6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112 not found: ID does not exist" containerID="6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.519306 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112"} err="failed to get container status \"6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112\": rpc error: code = NotFound desc = could not find container \"6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112\": container with ID starting with 6edafee502e0dcc750873e1e0c7ad8021a3e7c3a827cc02c812b19b70b002112 not found: ID does not exist" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.519320 4800 scope.go:117] "RemoveContainer" containerID="5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24" Jan 22 00:22:42 crc kubenswrapper[4800]: E0122 00:22:42.519840 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24\": container with ID starting with 5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24 not found: ID does not exist" containerID="5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.519954 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24"} err="failed to get container status \"5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24\": rpc error: code = NotFound desc = could not find container \"5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24\": container with ID starting with 5068467992cb54db79a24f97807631c6e28ca1a9ef7d6301c4d44ca04e933a24 not found: ID does not exist" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.520385 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc5xj\" (UniqueName: \"kubernetes.io/projected/da7c939b-1cc4-46a7-af60-968d3efbdcd9-kube-api-access-nc5xj\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:42 crc kubenswrapper[4800]: I0122 00:22:42.520406 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:43 crc kubenswrapper[4800]: I0122 00:22:43.408930 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da7c939b-1cc4-46a7-af60-968d3efbdcd9" (UID: "da7c939b-1cc4-46a7-af60-968d3efbdcd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:43 crc kubenswrapper[4800]: I0122 00:22:43.432741 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c939b-1cc4-46a7-af60-968d3efbdcd9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:43 crc kubenswrapper[4800]: I0122 00:22:43.693189 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:43 crc kubenswrapper[4800]: I0122 00:22:43.700115 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q48wn"] Jan 22 00:22:44 crc kubenswrapper[4800]: I0122 00:22:44.825182 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" path="/var/lib/kubelet/pods/da7c939b-1cc4-46a7-af60-968d3efbdcd9/volumes" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.925247 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:45 crc kubenswrapper[4800]: E0122 00:22:45.925508 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="extract-utilities" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.925520 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="extract-utilities" Jan 22 00:22:45 crc kubenswrapper[4800]: E0122 00:22:45.925538 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="extract-content" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.925545 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="extract-content" Jan 22 00:22:45 crc kubenswrapper[4800]: E0122 00:22:45.925561 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="registry-server" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.925570 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="registry-server" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.925694 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7c939b-1cc4-46a7-af60-968d3efbdcd9" containerName="registry-server" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.926695 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:45 crc kubenswrapper[4800]: I0122 00:22:45.934483 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.069499 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.069823 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cswl2\" (UniqueName: \"kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.069878 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.171646 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.171696 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cswl2\" (UniqueName: \"kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.171735 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.172171 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.172346 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.197540 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cswl2\" (UniqueName: \"kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2\") pod \"certified-operators-fnwlz\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.256860 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:46 crc kubenswrapper[4800]: I0122 00:22:46.765961 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:47 crc kubenswrapper[4800]: I0122 00:22:47.491341 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerStarted","Data":"a2fa0d9f14013e96020ab598b720ca22d83283b65060827ba2464e116e183fdb"} Jan 22 00:22:48 crc kubenswrapper[4800]: I0122 00:22:48.498926 4800 generic.go:334] "Generic (PLEG): container finished" podID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerID="0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab" exitCode=0 Jan 22 00:22:48 crc kubenswrapper[4800]: I0122 00:22:48.499002 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerDied","Data":"0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab"} Jan 22 00:22:49 crc kubenswrapper[4800]: I0122 00:22:49.452315 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:22:49 crc kubenswrapper[4800]: I0122 00:22:49.452635 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:22:50 crc kubenswrapper[4800]: I0122 00:22:50.515789 4800 generic.go:334] "Generic (PLEG): container finished" podID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerID="cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15" exitCode=0 Jan 22 00:22:50 crc kubenswrapper[4800]: I0122 00:22:50.515868 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerDied","Data":"cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15"} Jan 22 00:22:51 crc kubenswrapper[4800]: I0122 00:22:51.523949 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerStarted","Data":"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e"} Jan 22 00:22:51 crc kubenswrapper[4800]: I0122 00:22:51.552491 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fnwlz" podStartSLOduration=4.13723886 podStartE2EDuration="6.552464337s" podCreationTimestamp="2026-01-22 00:22:45 +0000 UTC" firstStartedPulling="2026-01-22 00:22:48.500414361 +0000 UTC m=+943.980717399" lastFinishedPulling="2026-01-22 00:22:50.915639838 +0000 UTC m=+946.395942876" observedRunningTime="2026-01-22 00:22:51.545284004 +0000 UTC m=+947.025587062" watchObservedRunningTime="2026-01-22 00:22:51.552464337 +0000 UTC m=+947.032767375" Jan 22 00:22:56 crc kubenswrapper[4800]: I0122 00:22:56.257842 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:56 crc kubenswrapper[4800]: I0122 00:22:56.259142 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:56 crc kubenswrapper[4800]: I0122 00:22:56.296949 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:56 crc kubenswrapper[4800]: I0122 00:22:56.592465 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:56 crc kubenswrapper[4800]: I0122 00:22:56.642298 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:58 crc kubenswrapper[4800]: I0122 00:22:58.566813 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fnwlz" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="registry-server" containerID="cri-o://ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e" gracePeriod=2 Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.425816 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.483509 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cswl2\" (UniqueName: \"kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2\") pod \"3212f5ad-47e2-467a-88c1-51418b9b4b92\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.483579 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content\") pod \"3212f5ad-47e2-467a-88c1-51418b9b4b92\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.483643 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities\") pod \"3212f5ad-47e2-467a-88c1-51418b9b4b92\" (UID: \"3212f5ad-47e2-467a-88c1-51418b9b4b92\") " Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.485921 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities" (OuterVolumeSpecName: "utilities") pod "3212f5ad-47e2-467a-88c1-51418b9b4b92" (UID: "3212f5ad-47e2-467a-88c1-51418b9b4b92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.498055 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2" (OuterVolumeSpecName: "kube-api-access-cswl2") pod "3212f5ad-47e2-467a-88c1-51418b9b4b92" (UID: "3212f5ad-47e2-467a-88c1-51418b9b4b92"). InnerVolumeSpecName "kube-api-access-cswl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.542519 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3212f5ad-47e2-467a-88c1-51418b9b4b92" (UID: "3212f5ad-47e2-467a-88c1-51418b9b4b92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.573558 4800 generic.go:334] "Generic (PLEG): container finished" podID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerID="ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e" exitCode=0 Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.573614 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerDied","Data":"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e"} Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.573650 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnwlz" event={"ID":"3212f5ad-47e2-467a-88c1-51418b9b4b92","Type":"ContainerDied","Data":"a2fa0d9f14013e96020ab598b720ca22d83283b65060827ba2464e116e183fdb"} Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.573659 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnwlz" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.573673 4800 scope.go:117] "RemoveContainer" containerID="ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.585475 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.585505 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cswl2\" (UniqueName: \"kubernetes.io/projected/3212f5ad-47e2-467a-88c1-51418b9b4b92-kube-api-access-cswl2\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.585516 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3212f5ad-47e2-467a-88c1-51418b9b4b92-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.590688 4800 scope.go:117] "RemoveContainer" containerID="cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.598298 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.604511 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fnwlz"] Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.621225 4800 scope.go:117] "RemoveContainer" containerID="0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.633645 4800 scope.go:117] "RemoveContainer" containerID="ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e" Jan 22 00:22:59 crc kubenswrapper[4800]: E0122 00:22:59.634060 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e\": container with ID starting with ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e not found: ID does not exist" containerID="ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.634115 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e"} err="failed to get container status \"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e\": rpc error: code = NotFound desc = could not find container \"ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e\": container with ID starting with ada8e2fbbcf2d9cf50277da5915daecc361277ddefcba448b312539b92a1777e not found: ID does not exist" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.634139 4800 scope.go:117] "RemoveContainer" containerID="cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15" Jan 22 00:22:59 crc kubenswrapper[4800]: E0122 00:22:59.634555 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15\": container with ID starting with cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15 not found: ID does not exist" containerID="cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.634604 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15"} err="failed to get container status \"cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15\": rpc error: code = NotFound desc = could not find container \"cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15\": container with ID starting with cb04992a8ae22400b9c4fa04592c076827baf6413413aab9020c4dda4f9dce15 not found: ID does not exist" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.634638 4800 scope.go:117] "RemoveContainer" containerID="0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab" Jan 22 00:22:59 crc kubenswrapper[4800]: E0122 00:22:59.635164 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab\": container with ID starting with 0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab not found: ID does not exist" containerID="0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab" Jan 22 00:22:59 crc kubenswrapper[4800]: I0122 00:22:59.635225 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab"} err="failed to get container status \"0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab\": rpc error: code = NotFound desc = could not find container \"0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab\": container with ID starting with 0b61586bfea7977b049d1e2a9becc11cffa1c52c1f5af8f98c69fa2104f26eab not found: ID does not exist" Jan 22 00:23:00 crc kubenswrapper[4800]: I0122 00:23:00.826473 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" path="/var/lib/kubelet/pods/3212f5ad-47e2-467a-88c1-51418b9b4b92/volumes" Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.451429 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.452041 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.452095 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.452817 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.452893 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea" gracePeriod=600 Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.692160 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea" exitCode=0 Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.692244 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea"} Jan 22 00:23:19 crc kubenswrapper[4800]: I0122 00:23:19.692435 4800 scope.go:117] "RemoveContainer" containerID="a87b5b44ea3bd201e448f6a3f72b915c46a0cac161305ec213de2e227780865c" Jan 22 00:23:20 crc kubenswrapper[4800]: I0122 00:23:20.702191 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef"} Jan 22 00:25:19 crc kubenswrapper[4800]: I0122 00:25:19.451702 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:25:19 crc kubenswrapper[4800]: I0122 00:25:19.452261 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:25:49 crc kubenswrapper[4800]: I0122 00:25:49.452032 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:25:49 crc kubenswrapper[4800]: I0122 00:25:49.452612 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.452485 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.453069 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.453119 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.453737 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.453798 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef" gracePeriod=600 Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.873990 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef" exitCode=0 Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.874186 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef"} Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.874267 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf"} Jan 22 00:26:19 crc kubenswrapper[4800]: I0122 00:26:19.874312 4800 scope.go:117] "RemoveContainer" containerID="536d4384d0fae0b450b545638a9bab78f606e9418e51e6f9bccaf55951be5cea" Jan 22 00:28:19 crc kubenswrapper[4800]: I0122 00:28:19.452016 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:28:19 crc kubenswrapper[4800]: I0122 00:28:19.452574 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:28:49 crc kubenswrapper[4800]: I0122 00:28:49.451801 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:28:49 crc kubenswrapper[4800]: I0122 00:28:49.452422 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:29:19 crc kubenswrapper[4800]: I0122 00:29:19.451932 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:29:19 crc kubenswrapper[4800]: I0122 00:29:19.452505 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:29:19 crc kubenswrapper[4800]: I0122 00:29:19.452834 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:29:19 crc kubenswrapper[4800]: I0122 00:29:19.453482 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:29:19 crc kubenswrapper[4800]: I0122 00:29:19.453537 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf" gracePeriod=600 Jan 22 00:29:20 crc kubenswrapper[4800]: I0122 00:29:20.542940 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf" exitCode=0 Jan 22 00:29:20 crc kubenswrapper[4800]: I0122 00:29:20.542949 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf"} Jan 22 00:29:20 crc kubenswrapper[4800]: I0122 00:29:20.543542 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f"} Jan 22 00:29:20 crc kubenswrapper[4800]: I0122 00:29:20.543568 4800 scope.go:117] "RemoveContainer" containerID="3aa890eafc48fe3ed6d37b89cbad34bc4dddb81117bd87b77e3b440fb720d6ef" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.151443 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x"] Jan 22 00:30:00 crc kubenswrapper[4800]: E0122 00:30:00.152273 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="extract-content" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.152291 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="extract-content" Jan 22 00:30:00 crc kubenswrapper[4800]: E0122 00:30:00.152313 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="extract-utilities" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.152322 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="extract-utilities" Jan 22 00:30:00 crc kubenswrapper[4800]: E0122 00:30:00.152615 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="registry-server" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.152629 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="registry-server" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.152751 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="3212f5ad-47e2-467a-88c1-51418b9b4b92" containerName="registry-server" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.153375 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.155557 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.157276 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.161949 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x"] Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.295525 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.295843 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.296014 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hbb\" (UniqueName: \"kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.397373 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.397438 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hbb\" (UniqueName: \"kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.397499 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.398376 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.416851 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.421428 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hbb\" (UniqueName: \"kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb\") pod \"collect-profiles-29484030-2292x\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.474879 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.662189 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x"] Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.832250 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" event={"ID":"0b7962c8-a1ef-4411-868a-b1cfb89cc82c","Type":"ContainerStarted","Data":"1dee2bebe559152710b99842b79d0b0920b34e1dbd180863894765eec709da8e"} Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.832304 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" event={"ID":"0b7962c8-a1ef-4411-868a-b1cfb89cc82c","Type":"ContainerStarted","Data":"003dfc3bc99a2221f54ab3a8796d4a27d702f634501bc2714dbf4041ba270e95"} Jan 22 00:30:00 crc kubenswrapper[4800]: I0122 00:30:00.849228 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" podStartSLOduration=0.849204069 podStartE2EDuration="849.204069ms" podCreationTimestamp="2026-01-22 00:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:30:00.847478703 +0000 UTC m=+1376.327781751" watchObservedRunningTime="2026-01-22 00:30:00.849204069 +0000 UTC m=+1376.329507107" Jan 22 00:30:01 crc kubenswrapper[4800]: I0122 00:30:01.841185 4800 generic.go:334] "Generic (PLEG): container finished" podID="0b7962c8-a1ef-4411-868a-b1cfb89cc82c" containerID="1dee2bebe559152710b99842b79d0b0920b34e1dbd180863894765eec709da8e" exitCode=0 Jan 22 00:30:01 crc kubenswrapper[4800]: I0122 00:30:01.841354 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" event={"ID":"0b7962c8-a1ef-4411-868a-b1cfb89cc82c","Type":"ContainerDied","Data":"1dee2bebe559152710b99842b79d0b0920b34e1dbd180863894765eec709da8e"} Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.069052 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.137821 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume\") pod \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.137954 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume\") pod \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.138030 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82hbb\" (UniqueName: \"kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb\") pod \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\" (UID: \"0b7962c8-a1ef-4411-868a-b1cfb89cc82c\") " Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.138626 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b7962c8-a1ef-4411-868a-b1cfb89cc82c" (UID: "0b7962c8-a1ef-4411-868a-b1cfb89cc82c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.142834 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb" (OuterVolumeSpecName: "kube-api-access-82hbb") pod "0b7962c8-a1ef-4411-868a-b1cfb89cc82c" (UID: "0b7962c8-a1ef-4411-868a-b1cfb89cc82c"). InnerVolumeSpecName "kube-api-access-82hbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.143031 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b7962c8-a1ef-4411-868a-b1cfb89cc82c" (UID: "0b7962c8-a1ef-4411-868a-b1cfb89cc82c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.240119 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82hbb\" (UniqueName: \"kubernetes.io/projected/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-kube-api-access-82hbb\") on node \"crc\" DevicePath \"\"" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.240158 4800 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.240171 4800 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b7962c8-a1ef-4411-868a-b1cfb89cc82c-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.856484 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" event={"ID":"0b7962c8-a1ef-4411-868a-b1cfb89cc82c","Type":"ContainerDied","Data":"003dfc3bc99a2221f54ab3a8796d4a27d702f634501bc2714dbf4041ba270e95"} Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.856560 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484030-2292x" Jan 22 00:30:03 crc kubenswrapper[4800]: I0122 00:30:03.856573 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="003dfc3bc99a2221f54ab3a8796d4a27d702f634501bc2714dbf4041ba270e95" Jan 22 00:31:19 crc kubenswrapper[4800]: I0122 00:31:19.451507 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:31:19 crc kubenswrapper[4800]: I0122 00:31:19.453744 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:31:49 crc kubenswrapper[4800]: I0122 00:31:49.452768 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:31:49 crc kubenswrapper[4800]: I0122 00:31:49.453490 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.452401 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.453376 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.453455 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.455682 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.455809 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" gracePeriod=600 Jan 22 00:32:19 crc kubenswrapper[4800]: E0122 00:32:19.589184 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.764983 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" exitCode=0 Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.765036 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f"} Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.765079 4800 scope.go:117] "RemoveContainer" containerID="a336ab5e0996ed0b75b22d77f9b68f3466167aff0c7a0db55e9ca4f8089b6caf" Jan 22 00:32:19 crc kubenswrapper[4800]: I0122 00:32:19.765467 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:32:19 crc kubenswrapper[4800]: E0122 00:32:19.765713 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:32:34 crc kubenswrapper[4800]: I0122 00:32:34.820970 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:32:34 crc kubenswrapper[4800]: E0122 00:32:34.821662 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:32:45 crc kubenswrapper[4800]: I0122 00:32:45.818727 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:32:45 crc kubenswrapper[4800]: E0122 00:32:45.819370 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:32:56 crc kubenswrapper[4800]: I0122 00:32:56.818469 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:32:56 crc kubenswrapper[4800]: E0122 00:32:56.819246 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:33:08 crc kubenswrapper[4800]: I0122 00:33:08.818372 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:33:08 crc kubenswrapper[4800]: E0122 00:33:08.819321 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.687295 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:18 crc kubenswrapper[4800]: E0122 00:33:18.688075 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7962c8-a1ef-4411-868a-b1cfb89cc82c" containerName="collect-profiles" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.688088 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7962c8-a1ef-4411-868a-b1cfb89cc82c" containerName="collect-profiles" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.688179 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7962c8-a1ef-4411-868a-b1cfb89cc82c" containerName="collect-profiles" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.688980 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.698142 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.742722 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.742768 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbbvj\" (UniqueName: \"kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.742812 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.844596 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.844975 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.845096 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbbvj\" (UniqueName: \"kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.845180 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.845495 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:18 crc kubenswrapper[4800]: I0122 00:33:18.869525 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbbvj\" (UniqueName: \"kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj\") pod \"certified-operators-rj8q9\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:19 crc kubenswrapper[4800]: I0122 00:33:19.017685 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:19 crc kubenswrapper[4800]: I0122 00:33:19.472091 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:19 crc kubenswrapper[4800]: I0122 00:33:19.818103 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:33:19 crc kubenswrapper[4800]: E0122 00:33:19.818408 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:33:20 crc kubenswrapper[4800]: I0122 00:33:20.171806 4800 generic.go:334] "Generic (PLEG): container finished" podID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerID="5969cfb5a8c0ff5e722392f2e91160a4bdf552d844d7c9a8adb1c0faf92efa3b" exitCode=0 Jan 22 00:33:20 crc kubenswrapper[4800]: I0122 00:33:20.171853 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerDied","Data":"5969cfb5a8c0ff5e722392f2e91160a4bdf552d844d7c9a8adb1c0faf92efa3b"} Jan 22 00:33:20 crc kubenswrapper[4800]: I0122 00:33:20.171902 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerStarted","Data":"9b23d8b3ebc3d913a3d399ab83f52b60396c0b4a09819be78fb284861cc794af"} Jan 22 00:33:20 crc kubenswrapper[4800]: I0122 00:33:20.173821 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:33:21 crc kubenswrapper[4800]: I0122 00:33:21.180858 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerStarted","Data":"829c93e94ded5166f6c4c1050c1d9201a7290f71523ac21ca9127399e68dd72c"} Jan 22 00:33:22 crc kubenswrapper[4800]: I0122 00:33:22.197072 4800 generic.go:334] "Generic (PLEG): container finished" podID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerID="829c93e94ded5166f6c4c1050c1d9201a7290f71523ac21ca9127399e68dd72c" exitCode=0 Jan 22 00:33:22 crc kubenswrapper[4800]: I0122 00:33:22.197128 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerDied","Data":"829c93e94ded5166f6c4c1050c1d9201a7290f71523ac21ca9127399e68dd72c"} Jan 22 00:33:23 crc kubenswrapper[4800]: I0122 00:33:23.203836 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerStarted","Data":"be7d952630737cd46a293bda38a2e0c9f8df4a46b3d62b72a552a455c7e20d0d"} Jan 22 00:33:23 crc kubenswrapper[4800]: I0122 00:33:23.230501 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rj8q9" podStartSLOduration=2.829519866 podStartE2EDuration="5.230466159s" podCreationTimestamp="2026-01-22 00:33:18 +0000 UTC" firstStartedPulling="2026-01-22 00:33:20.173590584 +0000 UTC m=+1575.653893622" lastFinishedPulling="2026-01-22 00:33:22.574536877 +0000 UTC m=+1578.054839915" observedRunningTime="2026-01-22 00:33:23.219578877 +0000 UTC m=+1578.699881915" watchObservedRunningTime="2026-01-22 00:33:23.230466159 +0000 UTC m=+1578.710769197" Jan 22 00:33:29 crc kubenswrapper[4800]: I0122 00:33:29.018943 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:29 crc kubenswrapper[4800]: I0122 00:33:29.019452 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:29 crc kubenswrapper[4800]: I0122 00:33:29.065084 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:29 crc kubenswrapper[4800]: I0122 00:33:29.324805 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:29 crc kubenswrapper[4800]: I0122 00:33:29.366913 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:31 crc kubenswrapper[4800]: I0122 00:33:31.248285 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rj8q9" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="registry-server" containerID="cri-o://be7d952630737cd46a293bda38a2e0c9f8df4a46b3d62b72a552a455c7e20d0d" gracePeriod=2 Jan 22 00:33:31 crc kubenswrapper[4800]: I0122 00:33:31.818055 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:33:31 crc kubenswrapper[4800]: E0122 00:33:31.818334 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.255372 4800 generic.go:334] "Generic (PLEG): container finished" podID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerID="be7d952630737cd46a293bda38a2e0c9f8df4a46b3d62b72a552a455c7e20d0d" exitCode=0 Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.255727 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerDied","Data":"be7d952630737cd46a293bda38a2e0c9f8df4a46b3d62b72a552a455c7e20d0d"} Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.255754 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj8q9" event={"ID":"75a54c6f-d0a7-4039-8316-6751bd0ce503","Type":"ContainerDied","Data":"9b23d8b3ebc3d913a3d399ab83f52b60396c0b4a09819be78fb284861cc794af"} Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.255765 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b23d8b3ebc3d913a3d399ab83f52b60396c0b4a09819be78fb284861cc794af" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.258078 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.340484 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities\") pod \"75a54c6f-d0a7-4039-8316-6751bd0ce503\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.340579 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbbvj\" (UniqueName: \"kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj\") pod \"75a54c6f-d0a7-4039-8316-6751bd0ce503\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.340624 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content\") pod \"75a54c6f-d0a7-4039-8316-6751bd0ce503\" (UID: \"75a54c6f-d0a7-4039-8316-6751bd0ce503\") " Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.341815 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities" (OuterVolumeSpecName: "utilities") pod "75a54c6f-d0a7-4039-8316-6751bd0ce503" (UID: "75a54c6f-d0a7-4039-8316-6751bd0ce503"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.347268 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj" (OuterVolumeSpecName: "kube-api-access-pbbvj") pod "75a54c6f-d0a7-4039-8316-6751bd0ce503" (UID: "75a54c6f-d0a7-4039-8316-6751bd0ce503"). InnerVolumeSpecName "kube-api-access-pbbvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.382230 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75a54c6f-d0a7-4039-8316-6751bd0ce503" (UID: "75a54c6f-d0a7-4039-8316-6751bd0ce503"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.442353 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.442393 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbbvj\" (UniqueName: \"kubernetes.io/projected/75a54c6f-d0a7-4039-8316-6751bd0ce503-kube-api-access-pbbvj\") on node \"crc\" DevicePath \"\"" Jan 22 00:33:32 crc kubenswrapper[4800]: I0122 00:33:32.442414 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a54c6f-d0a7-4039-8316-6751bd0ce503-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:33:33 crc kubenswrapper[4800]: I0122 00:33:33.260220 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj8q9" Jan 22 00:33:33 crc kubenswrapper[4800]: I0122 00:33:33.277875 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:33 crc kubenswrapper[4800]: I0122 00:33:33.281921 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rj8q9"] Jan 22 00:33:34 crc kubenswrapper[4800]: I0122 00:33:34.839471 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" path="/var/lib/kubelet/pods/75a54c6f-d0a7-4039-8316-6751bd0ce503/volumes" Jan 22 00:33:42 crc kubenswrapper[4800]: I0122 00:33:42.820057 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:33:42 crc kubenswrapper[4800]: E0122 00:33:42.821723 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:33:55 crc kubenswrapper[4800]: I0122 00:33:55.819090 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:33:55 crc kubenswrapper[4800]: E0122 00:33:55.819813 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:34:10 crc kubenswrapper[4800]: I0122 00:34:10.818622 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:34:10 crc kubenswrapper[4800]: E0122 00:34:10.819397 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:34:21 crc kubenswrapper[4800]: I0122 00:34:21.818723 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:34:21 crc kubenswrapper[4800]: E0122 00:34:21.819530 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:34:35 crc kubenswrapper[4800]: I0122 00:34:35.818424 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:34:35 crc kubenswrapper[4800]: E0122 00:34:35.819064 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:34:48 crc kubenswrapper[4800]: I0122 00:34:48.817754 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:34:48 crc kubenswrapper[4800]: E0122 00:34:48.818485 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:35:03 crc kubenswrapper[4800]: I0122 00:35:03.817972 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:35:03 crc kubenswrapper[4800]: E0122 00:35:03.818717 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:35:18 crc kubenswrapper[4800]: I0122 00:35:18.817818 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:35:18 crc kubenswrapper[4800]: E0122 00:35:18.818534 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:35:30 crc kubenswrapper[4800]: I0122 00:35:30.818598 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:35:30 crc kubenswrapper[4800]: E0122 00:35:30.819518 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:35:41 crc kubenswrapper[4800]: I0122 00:35:41.818693 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:35:41 crc kubenswrapper[4800]: E0122 00:35:41.820074 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:35:53 crc kubenswrapper[4800]: I0122 00:35:53.817975 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:35:53 crc kubenswrapper[4800]: E0122 00:35:53.818770 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:36:04 crc kubenswrapper[4800]: I0122 00:36:04.825740 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:36:04 crc kubenswrapper[4800]: E0122 00:36:04.826527 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:36:16 crc kubenswrapper[4800]: I0122 00:36:16.818532 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:36:16 crc kubenswrapper[4800]: E0122 00:36:16.819367 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:36:30 crc kubenswrapper[4800]: I0122 00:36:30.823483 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:36:30 crc kubenswrapper[4800]: E0122 00:36:30.824381 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:36:41 crc kubenswrapper[4800]: I0122 00:36:41.818625 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:36:41 crc kubenswrapper[4800]: E0122 00:36:41.826502 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:36:54 crc kubenswrapper[4800]: I0122 00:36:54.820181 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:36:54 crc kubenswrapper[4800]: E0122 00:36:54.820984 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:37:08 crc kubenswrapper[4800]: I0122 00:37:08.818053 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:37:08 crc kubenswrapper[4800]: E0122 00:37:08.819084 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:37:20 crc kubenswrapper[4800]: I0122 00:37:20.819180 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:37:21 crc kubenswrapper[4800]: I0122 00:37:21.824871 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1"} Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.375953 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:40 crc kubenswrapper[4800]: E0122 00:38:40.376910 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="extract-utilities" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.376929 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="extract-utilities" Jan 22 00:38:40 crc kubenswrapper[4800]: E0122 00:38:40.376956 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="extract-content" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.376963 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="extract-content" Jan 22 00:38:40 crc kubenswrapper[4800]: E0122 00:38:40.376977 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="registry-server" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.376984 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="registry-server" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.377106 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a54c6f-d0a7-4039-8316-6751bd0ce503" containerName="registry-server" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.379118 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.386631 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.512125 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kclqw\" (UniqueName: \"kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.512213 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.512256 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.614193 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.614317 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kclqw\" (UniqueName: \"kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.614377 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.614860 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.615205 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.636655 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kclqw\" (UniqueName: \"kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw\") pod \"redhat-operators-29hsw\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:40 crc kubenswrapper[4800]: I0122 00:38:40.741922 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:41 crc kubenswrapper[4800]: I0122 00:38:41.182683 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:41 crc kubenswrapper[4800]: W0122 00:38:41.207363 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8653f0c_6cfb_42b6_8305_0df21f3a76cf.slice/crio-e7a7761be1e4d14b1e2d21f6de9688c4e754d155974bb72d52789cd1a697acfc WatchSource:0}: Error finding container e7a7761be1e4d14b1e2d21f6de9688c4e754d155974bb72d52789cd1a697acfc: Status 404 returned error can't find the container with id e7a7761be1e4d14b1e2d21f6de9688c4e754d155974bb72d52789cd1a697acfc Jan 22 00:38:41 crc kubenswrapper[4800]: I0122 00:38:41.345807 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerStarted","Data":"e7a7761be1e4d14b1e2d21f6de9688c4e754d155974bb72d52789cd1a697acfc"} Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.169259 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.170417 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.188406 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.238062 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wghg9\" (UniqueName: \"kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.238146 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.238192 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.339585 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.339677 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wghg9\" (UniqueName: \"kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.339708 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.340198 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.340414 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.351765 4800 generic.go:334] "Generic (PLEG): container finished" podID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerID="017ce2f33ea51cde70f8cf0a9dba91ce86dc787af8bfcb49fa14a6decee26814" exitCode=0 Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.351823 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerDied","Data":"017ce2f33ea51cde70f8cf0a9dba91ce86dc787af8bfcb49fa14a6decee26814"} Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.353242 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.364512 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wghg9\" (UniqueName: \"kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9\") pod \"community-operators-z67dv\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.484650 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:42 crc kubenswrapper[4800]: W0122 00:38:42.732727 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30d60a88_8897_49eb_b9b6_5a80e4a82cb3.slice/crio-abfb332d663de8e542b21b0bfc99f9775117c5704b76c20b5d752b5c9b18e567 WatchSource:0}: Error finding container abfb332d663de8e542b21b0bfc99f9775117c5704b76c20b5d752b5c9b18e567: Status 404 returned error can't find the container with id abfb332d663de8e542b21b0bfc99f9775117c5704b76c20b5d752b5c9b18e567 Jan 22 00:38:42 crc kubenswrapper[4800]: I0122 00:38:42.733141 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:43 crc kubenswrapper[4800]: I0122 00:38:43.360490 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerStarted","Data":"c2697c2dfed58bd538884ec56b0c7e760b44f86243b90a8e17f971191ca45752"} Jan 22 00:38:43 crc kubenswrapper[4800]: I0122 00:38:43.362251 4800 generic.go:334] "Generic (PLEG): container finished" podID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerID="9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081" exitCode=0 Jan 22 00:38:43 crc kubenswrapper[4800]: I0122 00:38:43.362309 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerDied","Data":"9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081"} Jan 22 00:38:43 crc kubenswrapper[4800]: I0122 00:38:43.362347 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerStarted","Data":"abfb332d663de8e542b21b0bfc99f9775117c5704b76c20b5d752b5c9b18e567"} Jan 22 00:38:44 crc kubenswrapper[4800]: I0122 00:38:44.369931 4800 generic.go:334] "Generic (PLEG): container finished" podID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerID="c2697c2dfed58bd538884ec56b0c7e760b44f86243b90a8e17f971191ca45752" exitCode=0 Jan 22 00:38:44 crc kubenswrapper[4800]: I0122 00:38:44.369986 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerDied","Data":"c2697c2dfed58bd538884ec56b0c7e760b44f86243b90a8e17f971191ca45752"} Jan 22 00:38:45 crc kubenswrapper[4800]: I0122 00:38:45.376312 4800 generic.go:334] "Generic (PLEG): container finished" podID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerID="ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985" exitCode=0 Jan 22 00:38:45 crc kubenswrapper[4800]: I0122 00:38:45.376380 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerDied","Data":"ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985"} Jan 22 00:38:45 crc kubenswrapper[4800]: I0122 00:38:45.380751 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerStarted","Data":"744d508f14071e5568c4c35546f8b362daaf94e91c6944ee6a5d7af957582b2d"} Jan 22 00:38:45 crc kubenswrapper[4800]: I0122 00:38:45.413832 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-29hsw" podStartSLOduration=2.976749158 podStartE2EDuration="5.413809021s" podCreationTimestamp="2026-01-22 00:38:40 +0000 UTC" firstStartedPulling="2026-01-22 00:38:42.353067017 +0000 UTC m=+1897.833370055" lastFinishedPulling="2026-01-22 00:38:44.79012687 +0000 UTC m=+1900.270429918" observedRunningTime="2026-01-22 00:38:45.41153234 +0000 UTC m=+1900.891835378" watchObservedRunningTime="2026-01-22 00:38:45.413809021 +0000 UTC m=+1900.894112059" Jan 22 00:38:46 crc kubenswrapper[4800]: I0122 00:38:46.391575 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerStarted","Data":"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80"} Jan 22 00:38:46 crc kubenswrapper[4800]: I0122 00:38:46.411563 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z67dv" podStartSLOduration=2.001713321 podStartE2EDuration="4.411542664s" podCreationTimestamp="2026-01-22 00:38:42 +0000 UTC" firstStartedPulling="2026-01-22 00:38:43.363769429 +0000 UTC m=+1898.844072467" lastFinishedPulling="2026-01-22 00:38:45.773598772 +0000 UTC m=+1901.253901810" observedRunningTime="2026-01-22 00:38:46.405564694 +0000 UTC m=+1901.885867752" watchObservedRunningTime="2026-01-22 00:38:46.411542664 +0000 UTC m=+1901.891845702" Jan 22 00:38:50 crc kubenswrapper[4800]: I0122 00:38:50.743529 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:50 crc kubenswrapper[4800]: I0122 00:38:50.743910 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:50 crc kubenswrapper[4800]: I0122 00:38:50.826464 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:51 crc kubenswrapper[4800]: I0122 00:38:51.462160 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:51 crc kubenswrapper[4800]: I0122 00:38:51.560827 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:52 crc kubenswrapper[4800]: I0122 00:38:52.485494 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:52 crc kubenswrapper[4800]: I0122 00:38:52.485573 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:52 crc kubenswrapper[4800]: I0122 00:38:52.540311 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:53 crc kubenswrapper[4800]: I0122 00:38:53.433994 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-29hsw" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="registry-server" containerID="cri-o://744d508f14071e5568c4c35546f8b362daaf94e91c6944ee6a5d7af957582b2d" gracePeriod=2 Jan 22 00:38:53 crc kubenswrapper[4800]: I0122 00:38:53.485031 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:53 crc kubenswrapper[4800]: I0122 00:38:53.967735 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.449420 4800 generic.go:334] "Generic (PLEG): container finished" podID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerID="744d508f14071e5568c4c35546f8b362daaf94e91c6944ee6a5d7af957582b2d" exitCode=0 Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.449659 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z67dv" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="registry-server" containerID="cri-o://1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80" gracePeriod=2 Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.449729 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerDied","Data":"744d508f14071e5568c4c35546f8b362daaf94e91c6944ee6a5d7af957582b2d"} Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.640735 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.746788 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kclqw\" (UniqueName: \"kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw\") pod \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.746864 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities\") pod \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.746882 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content\") pod \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\" (UID: \"e8653f0c-6cfb-42b6-8305-0df21f3a76cf\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.747787 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities" (OuterVolumeSpecName: "utilities") pod "e8653f0c-6cfb-42b6-8305-0df21f3a76cf" (UID: "e8653f0c-6cfb-42b6-8305-0df21f3a76cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.754482 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw" (OuterVolumeSpecName: "kube-api-access-kclqw") pod "e8653f0c-6cfb-42b6-8305-0df21f3a76cf" (UID: "e8653f0c-6cfb-42b6-8305-0df21f3a76cf"). InnerVolumeSpecName "kube-api-access-kclqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.812756 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.848472 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kclqw\" (UniqueName: \"kubernetes.io/projected/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-kube-api-access-kclqw\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.848503 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.867802 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8653f0c-6cfb-42b6-8305-0df21f3a76cf" (UID: "e8653f0c-6cfb-42b6-8305-0df21f3a76cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.950012 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content\") pod \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.950060 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities\") pod \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.950181 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wghg9\" (UniqueName: \"kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9\") pod \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\" (UID: \"30d60a88-8897-49eb-b9b6-5a80e4a82cb3\") " Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.950485 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8653f0c-6cfb-42b6-8305-0df21f3a76cf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.965483 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities" (OuterVolumeSpecName: "utilities") pod "30d60a88-8897-49eb-b9b6-5a80e4a82cb3" (UID: "30d60a88-8897-49eb-b9b6-5a80e4a82cb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.968430 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9" (OuterVolumeSpecName: "kube-api-access-wghg9") pod "30d60a88-8897-49eb-b9b6-5a80e4a82cb3" (UID: "30d60a88-8897-49eb-b9b6-5a80e4a82cb3"). InnerVolumeSpecName "kube-api-access-wghg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:38:55 crc kubenswrapper[4800]: I0122 00:38:55.998689 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30d60a88-8897-49eb-b9b6-5a80e4a82cb3" (UID: "30d60a88-8897-49eb-b9b6-5a80e4a82cb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.051521 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.051583 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.051960 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wghg9\" (UniqueName: \"kubernetes.io/projected/30d60a88-8897-49eb-b9b6-5a80e4a82cb3-kube-api-access-wghg9\") on node \"crc\" DevicePath \"\"" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.459290 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-29hsw" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.459323 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-29hsw" event={"ID":"e8653f0c-6cfb-42b6-8305-0df21f3a76cf","Type":"ContainerDied","Data":"e7a7761be1e4d14b1e2d21f6de9688c4e754d155974bb72d52789cd1a697acfc"} Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.459428 4800 scope.go:117] "RemoveContainer" containerID="744d508f14071e5568c4c35546f8b362daaf94e91c6944ee6a5d7af957582b2d" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.466757 4800 generic.go:334] "Generic (PLEG): container finished" podID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerID="1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80" exitCode=0 Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.466818 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67dv" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.466850 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerDied","Data":"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80"} Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.467360 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67dv" event={"ID":"30d60a88-8897-49eb-b9b6-5a80e4a82cb3","Type":"ContainerDied","Data":"abfb332d663de8e542b21b0bfc99f9775117c5704b76c20b5d752b5c9b18e567"} Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.494678 4800 scope.go:117] "RemoveContainer" containerID="c2697c2dfed58bd538884ec56b0c7e760b44f86243b90a8e17f971191ca45752" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.495287 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.500916 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-29hsw"] Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.516765 4800 scope.go:117] "RemoveContainer" containerID="017ce2f33ea51cde70f8cf0a9dba91ce86dc787af8bfcb49fa14a6decee26814" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.539400 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.542786 4800 scope.go:117] "RemoveContainer" containerID="1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.546033 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z67dv"] Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.558477 4800 scope.go:117] "RemoveContainer" containerID="ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.591292 4800 scope.go:117] "RemoveContainer" containerID="9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.608664 4800 scope.go:117] "RemoveContainer" containerID="1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80" Jan 22 00:38:56 crc kubenswrapper[4800]: E0122 00:38:56.609153 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80\": container with ID starting with 1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80 not found: ID does not exist" containerID="1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.609196 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80"} err="failed to get container status \"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80\": rpc error: code = NotFound desc = could not find container \"1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80\": container with ID starting with 1a7117d7f077a49cdc1af39280bea3dffd58de6c3010f51cc0d95fe17d2ede80 not found: ID does not exist" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.609227 4800 scope.go:117] "RemoveContainer" containerID="ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985" Jan 22 00:38:56 crc kubenswrapper[4800]: E0122 00:38:56.609553 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985\": container with ID starting with ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985 not found: ID does not exist" containerID="ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.609589 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985"} err="failed to get container status \"ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985\": rpc error: code = NotFound desc = could not find container \"ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985\": container with ID starting with ebf64974f8e660dc188d9c23925e24e0f2bac119d91d893a0c897cc54359e985 not found: ID does not exist" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.609608 4800 scope.go:117] "RemoveContainer" containerID="9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081" Jan 22 00:38:56 crc kubenswrapper[4800]: E0122 00:38:56.609805 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081\": container with ID starting with 9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081 not found: ID does not exist" containerID="9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.609833 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081"} err="failed to get container status \"9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081\": rpc error: code = NotFound desc = could not find container \"9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081\": container with ID starting with 9025ef295ec7782e1aeb2ed9d060e059bd5e21fcfea9b7bc4dce3a40debbd081 not found: ID does not exist" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.826311 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" path="/var/lib/kubelet/pods/30d60a88-8897-49eb-b9b6-5a80e4a82cb3/volumes" Jan 22 00:38:56 crc kubenswrapper[4800]: I0122 00:38:56.827171 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" path="/var/lib/kubelet/pods/e8653f0c-6cfb-42b6-8305-0df21f3a76cf/volumes" Jan 22 00:39:49 crc kubenswrapper[4800]: I0122 00:39:49.451368 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:39:49 crc kubenswrapper[4800]: I0122 00:39:49.452018 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:40:07 crc kubenswrapper[4800]: I0122 00:40:07.700183 4800 scope.go:117] "RemoveContainer" containerID="829c93e94ded5166f6c4c1050c1d9201a7290f71523ac21ca9127399e68dd72c" Jan 22 00:40:07 crc kubenswrapper[4800]: I0122 00:40:07.738381 4800 scope.go:117] "RemoveContainer" containerID="be7d952630737cd46a293bda38a2e0c9f8df4a46b3d62b72a552a455c7e20d0d" Jan 22 00:40:07 crc kubenswrapper[4800]: I0122 00:40:07.756480 4800 scope.go:117] "RemoveContainer" containerID="5969cfb5a8c0ff5e722392f2e91160a4bdf552d844d7c9a8adb1c0faf92efa3b" Jan 22 00:40:19 crc kubenswrapper[4800]: I0122 00:40:19.451466 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:40:19 crc kubenswrapper[4800]: I0122 00:40:19.452175 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:40:49 crc kubenswrapper[4800]: I0122 00:40:49.451554 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:40:49 crc kubenswrapper[4800]: I0122 00:40:49.452441 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:40:49 crc kubenswrapper[4800]: I0122 00:40:49.452513 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:40:49 crc kubenswrapper[4800]: I0122 00:40:49.453497 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:40:49 crc kubenswrapper[4800]: I0122 00:40:49.453590 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1" gracePeriod=600 Jan 22 00:40:50 crc kubenswrapper[4800]: I0122 00:40:50.224424 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1" exitCode=0 Jan 22 00:40:50 crc kubenswrapper[4800]: I0122 00:40:50.224908 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1"} Jan 22 00:40:50 crc kubenswrapper[4800]: I0122 00:40:50.224942 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2"} Jan 22 00:40:50 crc kubenswrapper[4800]: I0122 00:40:50.224960 4800 scope.go:117] "RemoveContainer" containerID="65a25cd63adecb6b40da7b13a3b291375e87bf185610d812c0960ac8d7ea528f" Jan 22 00:42:49 crc kubenswrapper[4800]: I0122 00:42:49.451607 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:42:49 crc kubenswrapper[4800]: I0122 00:42:49.452105 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:43:19 crc kubenswrapper[4800]: I0122 00:43:19.452081 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:43:19 crc kubenswrapper[4800]: I0122 00:43:19.452708 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:43:49 crc kubenswrapper[4800]: I0122 00:43:49.451989 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:43:49 crc kubenswrapper[4800]: I0122 00:43:49.452704 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:43:49 crc kubenswrapper[4800]: I0122 00:43:49.452754 4800 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" Jan 22 00:43:49 crc kubenswrapper[4800]: I0122 00:43:49.453408 4800 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2"} pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 00:43:49 crc kubenswrapper[4800]: I0122 00:43:49.453466 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" containerID="cri-o://84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" gracePeriod=600 Jan 22 00:43:49 crc kubenswrapper[4800]: E0122 00:43:49.571497 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:43:50 crc kubenswrapper[4800]: I0122 00:43:50.414857 4800 generic.go:334] "Generic (PLEG): container finished" podID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" exitCode=0 Jan 22 00:43:50 crc kubenswrapper[4800]: I0122 00:43:50.414961 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerDied","Data":"84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2"} Jan 22 00:43:50 crc kubenswrapper[4800]: I0122 00:43:50.415294 4800 scope.go:117] "RemoveContainer" containerID="aa215f1b067f82d7ce3da193a452c667aab5a71180588d0fa876b765088aa5e1" Jan 22 00:43:50 crc kubenswrapper[4800]: I0122 00:43:50.416329 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:43:50 crc kubenswrapper[4800]: E0122 00:43:50.416825 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:44:02 crc kubenswrapper[4800]: I0122 00:44:02.821729 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:44:02 crc kubenswrapper[4800]: E0122 00:44:02.822643 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871145 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871535 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871554 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871574 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871586 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871602 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="extract-utilities" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871614 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="extract-utilities" Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871635 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="extract-utilities" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871646 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="extract-utilities" Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871664 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="extract-content" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871675 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="extract-content" Jan 22 00:44:03 crc kubenswrapper[4800]: E0122 00:44:03.871689 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="extract-content" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871700 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="extract-content" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871927 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8653f0c-6cfb-42b6-8305-0df21f3a76cf" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.871951 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d60a88-8897-49eb-b9b6-5a80e4a82cb3" containerName="registry-server" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.873278 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.883731 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqxnt\" (UniqueName: \"kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.883858 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.883985 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.890482 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.986647 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.986728 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.986810 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqxnt\" (UniqueName: \"kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.987169 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:03 crc kubenswrapper[4800]: I0122 00:44:03.987291 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:04 crc kubenswrapper[4800]: I0122 00:44:04.019934 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqxnt\" (UniqueName: \"kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt\") pod \"certified-operators-dlt24\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:04 crc kubenswrapper[4800]: I0122 00:44:04.213042 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:04 crc kubenswrapper[4800]: I0122 00:44:04.461926 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:04 crc kubenswrapper[4800]: I0122 00:44:04.511470 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerStarted","Data":"216119f181826886148bb4b6c7a15898d9f8ac4963eb38570bcf667efef10979"} Jan 22 00:44:05 crc kubenswrapper[4800]: I0122 00:44:05.518296 4800 generic.go:334] "Generic (PLEG): container finished" podID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerID="95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538" exitCode=0 Jan 22 00:44:05 crc kubenswrapper[4800]: I0122 00:44:05.518346 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerDied","Data":"95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538"} Jan 22 00:44:05 crc kubenswrapper[4800]: I0122 00:44:05.523037 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:44:06 crc kubenswrapper[4800]: I0122 00:44:06.525354 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerStarted","Data":"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a"} Jan 22 00:44:07 crc kubenswrapper[4800]: I0122 00:44:07.531989 4800 generic.go:334] "Generic (PLEG): container finished" podID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerID="e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a" exitCode=0 Jan 22 00:44:07 crc kubenswrapper[4800]: I0122 00:44:07.532299 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerDied","Data":"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a"} Jan 22 00:44:08 crc kubenswrapper[4800]: I0122 00:44:08.549177 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerStarted","Data":"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c"} Jan 22 00:44:08 crc kubenswrapper[4800]: I0122 00:44:08.570444 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dlt24" podStartSLOduration=3.170928291 podStartE2EDuration="5.57042401s" podCreationTimestamp="2026-01-22 00:44:03 +0000 UTC" firstStartedPulling="2026-01-22 00:44:05.522838325 +0000 UTC m=+2221.003141363" lastFinishedPulling="2026-01-22 00:44:07.922334044 +0000 UTC m=+2223.402637082" observedRunningTime="2026-01-22 00:44:08.5689111 +0000 UTC m=+2224.049214148" watchObservedRunningTime="2026-01-22 00:44:08.57042401 +0000 UTC m=+2224.050727048" Jan 22 00:44:14 crc kubenswrapper[4800]: I0122 00:44:14.214216 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:14 crc kubenswrapper[4800]: I0122 00:44:14.214977 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:14 crc kubenswrapper[4800]: I0122 00:44:14.288097 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:14 crc kubenswrapper[4800]: I0122 00:44:14.636577 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:14 crc kubenswrapper[4800]: I0122 00:44:14.693157 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:16 crc kubenswrapper[4800]: I0122 00:44:16.607139 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dlt24" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="registry-server" containerID="cri-o://a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c" gracePeriod=2 Jan 22 00:44:16 crc kubenswrapper[4800]: I0122 00:44:16.819336 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:44:16 crc kubenswrapper[4800]: E0122 00:44:16.819855 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.043081 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.066502 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqxnt\" (UniqueName: \"kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt\") pod \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.066566 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities\") pod \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.066646 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content\") pod \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\" (UID: \"ba60685b-1acf-48ac-9a57-4ea6ad64c77b\") " Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.067923 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities" (OuterVolumeSpecName: "utilities") pod "ba60685b-1acf-48ac-9a57-4ea6ad64c77b" (UID: "ba60685b-1acf-48ac-9a57-4ea6ad64c77b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.080250 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt" (OuterVolumeSpecName: "kube-api-access-wqxnt") pod "ba60685b-1acf-48ac-9a57-4ea6ad64c77b" (UID: "ba60685b-1acf-48ac-9a57-4ea6ad64c77b"). InnerVolumeSpecName "kube-api-access-wqxnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.115631 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba60685b-1acf-48ac-9a57-4ea6ad64c77b" (UID: "ba60685b-1acf-48ac-9a57-4ea6ad64c77b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.168168 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.168193 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqxnt\" (UniqueName: \"kubernetes.io/projected/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-kube-api-access-wqxnt\") on node \"crc\" DevicePath \"\"" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.168204 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60685b-1acf-48ac-9a57-4ea6ad64c77b-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.613259 4800 generic.go:334] "Generic (PLEG): container finished" podID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerID="a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c" exitCode=0 Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.613303 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerDied","Data":"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c"} Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.613327 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlt24" event={"ID":"ba60685b-1acf-48ac-9a57-4ea6ad64c77b","Type":"ContainerDied","Data":"216119f181826886148bb4b6c7a15898d9f8ac4963eb38570bcf667efef10979"} Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.613342 4800 scope.go:117] "RemoveContainer" containerID="a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.613445 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlt24" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.631550 4800 scope.go:117] "RemoveContainer" containerID="e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.653381 4800 scope.go:117] "RemoveContainer" containerID="95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.684856 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.688823 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dlt24"] Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.691623 4800 scope.go:117] "RemoveContainer" containerID="a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c" Jan 22 00:44:17 crc kubenswrapper[4800]: E0122 00:44:17.692137 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c\": container with ID starting with a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c not found: ID does not exist" containerID="a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.692179 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c"} err="failed to get container status \"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c\": rpc error: code = NotFound desc = could not find container \"a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c\": container with ID starting with a92978eb9b51cfc24d431ff2567732bb12dcf1f2ddb9939d346deba115d9ab0c not found: ID does not exist" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.692204 4800 scope.go:117] "RemoveContainer" containerID="e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a" Jan 22 00:44:17 crc kubenswrapper[4800]: E0122 00:44:17.692674 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a\": container with ID starting with e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a not found: ID does not exist" containerID="e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.692700 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a"} err="failed to get container status \"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a\": rpc error: code = NotFound desc = could not find container \"e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a\": container with ID starting with e991f71ebc699bb3ce858a60824254fd8eaeec6c59aedcd9edd9eaccbf325f4a not found: ID does not exist" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.692713 4800 scope.go:117] "RemoveContainer" containerID="95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538" Jan 22 00:44:17 crc kubenswrapper[4800]: E0122 00:44:17.693015 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538\": container with ID starting with 95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538 not found: ID does not exist" containerID="95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538" Jan 22 00:44:17 crc kubenswrapper[4800]: I0122 00:44:17.693040 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538"} err="failed to get container status \"95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538\": rpc error: code = NotFound desc = could not find container \"95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538\": container with ID starting with 95b103648a0a993b2e349dc8dc408334e97ba9d2b9ee20aac602c8dbc4fa4538 not found: ID does not exist" Jan 22 00:44:18 crc kubenswrapper[4800]: I0122 00:44:18.830809 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" path="/var/lib/kubelet/pods/ba60685b-1acf-48ac-9a57-4ea6ad64c77b/volumes" Jan 22 00:44:27 crc kubenswrapper[4800]: I0122 00:44:27.819496 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:44:27 crc kubenswrapper[4800]: E0122 00:44:27.821329 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:44:38 crc kubenswrapper[4800]: I0122 00:44:38.821197 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:44:38 crc kubenswrapper[4800]: E0122 00:44:38.821913 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:44:52 crc kubenswrapper[4800]: I0122 00:44:52.818216 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:44:52 crc kubenswrapper[4800]: E0122 00:44:52.819107 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.139825 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf"] Jan 22 00:45:00 crc kubenswrapper[4800]: E0122 00:45:00.140850 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="extract-content" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.140866 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="extract-content" Jan 22 00:45:00 crc kubenswrapper[4800]: E0122 00:45:00.140913 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="registry-server" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.140921 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="registry-server" Jan 22 00:45:00 crc kubenswrapper[4800]: E0122 00:45:00.140967 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="extract-utilities" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.140975 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="extract-utilities" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.141118 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba60685b-1acf-48ac-9a57-4ea6ad64c77b" containerName="registry-server" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.141636 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.143523 4800 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.143753 4800 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.153722 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf"] Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.248904 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpq5t\" (UniqueName: \"kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.248984 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.249012 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.350515 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.350581 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.350626 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpq5t\" (UniqueName: \"kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.351676 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.363689 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.367057 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpq5t\" (UniqueName: \"kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t\") pod \"collect-profiles-29484045-v8vdf\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.461600 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.647696 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf"] Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.889508 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" event={"ID":"3f899bc4-42c1-4444-aafb-030278227777","Type":"ContainerStarted","Data":"3b58d88cb83fdf7e8684b9f45ddffcf7c8dd8f8952063560fffaa8b51ba391a5"} Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.889812 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" event={"ID":"3f899bc4-42c1-4444-aafb-030278227777","Type":"ContainerStarted","Data":"08b15d6d199d641ccde0364b3fda328291e78a59e2016a9f9c366acbbc9a8275"} Jan 22 00:45:00 crc kubenswrapper[4800]: I0122 00:45:00.907092 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" podStartSLOduration=0.907071705 podStartE2EDuration="907.071705ms" podCreationTimestamp="2026-01-22 00:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 00:45:00.907042274 +0000 UTC m=+2276.387345312" watchObservedRunningTime="2026-01-22 00:45:00.907071705 +0000 UTC m=+2276.387374743" Jan 22 00:45:01 crc kubenswrapper[4800]: I0122 00:45:01.896973 4800 generic.go:334] "Generic (PLEG): container finished" podID="3f899bc4-42c1-4444-aafb-030278227777" containerID="3b58d88cb83fdf7e8684b9f45ddffcf7c8dd8f8952063560fffaa8b51ba391a5" exitCode=0 Jan 22 00:45:01 crc kubenswrapper[4800]: I0122 00:45:01.897013 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" event={"ID":"3f899bc4-42c1-4444-aafb-030278227777","Type":"ContainerDied","Data":"3b58d88cb83fdf7e8684b9f45ddffcf7c8dd8f8952063560fffaa8b51ba391a5"} Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.170703 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.287476 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume\") pod \"3f899bc4-42c1-4444-aafb-030278227777\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.287644 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpq5t\" (UniqueName: \"kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t\") pod \"3f899bc4-42c1-4444-aafb-030278227777\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.287718 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume\") pod \"3f899bc4-42c1-4444-aafb-030278227777\" (UID: \"3f899bc4-42c1-4444-aafb-030278227777\") " Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.288475 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume" (OuterVolumeSpecName: "config-volume") pod "3f899bc4-42c1-4444-aafb-030278227777" (UID: "3f899bc4-42c1-4444-aafb-030278227777"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.288838 4800 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f899bc4-42c1-4444-aafb-030278227777-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.292810 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3f899bc4-42c1-4444-aafb-030278227777" (UID: "3f899bc4-42c1-4444-aafb-030278227777"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.293254 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t" (OuterVolumeSpecName: "kube-api-access-bpq5t") pod "3f899bc4-42c1-4444-aafb-030278227777" (UID: "3f899bc4-42c1-4444-aafb-030278227777"). InnerVolumeSpecName "kube-api-access-bpq5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.390213 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpq5t\" (UniqueName: \"kubernetes.io/projected/3f899bc4-42c1-4444-aafb-030278227777-kube-api-access-bpq5t\") on node \"crc\" DevicePath \"\"" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.390415 4800 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f899bc4-42c1-4444-aafb-030278227777-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.910258 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" event={"ID":"3f899bc4-42c1-4444-aafb-030278227777","Type":"ContainerDied","Data":"08b15d6d199d641ccde0364b3fda328291e78a59e2016a9f9c366acbbc9a8275"} Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.910300 4800 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08b15d6d199d641ccde0364b3fda328291e78a59e2016a9f9c366acbbc9a8275" Jan 22 00:45:03 crc kubenswrapper[4800]: I0122 00:45:03.910388 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484045-v8vdf" Jan 22 00:45:04 crc kubenswrapper[4800]: I0122 00:45:04.241945 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f"] Jan 22 00:45:04 crc kubenswrapper[4800]: I0122 00:45:04.249290 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484000-vjj9f"] Jan 22 00:45:04 crc kubenswrapper[4800]: I0122 00:45:04.836383 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f5bfc30-ad0a-486a-baad-9dc8249f7170" path="/var/lib/kubelet/pods/0f5bfc30-ad0a-486a-baad-9dc8249f7170/volumes" Jan 22 00:45:07 crc kubenswrapper[4800]: I0122 00:45:07.818686 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:45:07 crc kubenswrapper[4800]: E0122 00:45:07.819493 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:45:07 crc kubenswrapper[4800]: I0122 00:45:07.859047 4800 scope.go:117] "RemoveContainer" containerID="75627e11053cb23ff4dacefd3832ce37a3419f699288840153f5a35807fe7040" Jan 22 00:45:20 crc kubenswrapper[4800]: I0122 00:45:20.818564 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:45:20 crc kubenswrapper[4800]: E0122 00:45:20.819481 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:45:33 crc kubenswrapper[4800]: I0122 00:45:33.818626 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:45:33 crc kubenswrapper[4800]: E0122 00:45:33.819326 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:45:48 crc kubenswrapper[4800]: I0122 00:45:48.819860 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:45:48 crc kubenswrapper[4800]: E0122 00:45:48.820610 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:46:00 crc kubenswrapper[4800]: I0122 00:46:00.817988 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:46:00 crc kubenswrapper[4800]: E0122 00:46:00.818877 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:46:11 crc kubenswrapper[4800]: I0122 00:46:11.819327 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:46:11 crc kubenswrapper[4800]: E0122 00:46:11.819999 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:46:23 crc kubenswrapper[4800]: I0122 00:46:23.818491 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:46:23 crc kubenswrapper[4800]: E0122 00:46:23.819316 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:46:36 crc kubenswrapper[4800]: I0122 00:46:36.817779 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:46:36 crc kubenswrapper[4800]: E0122 00:46:36.818606 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:46:51 crc kubenswrapper[4800]: I0122 00:46:51.817732 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:46:51 crc kubenswrapper[4800]: E0122 00:46:51.818514 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:47:04 crc kubenswrapper[4800]: I0122 00:47:04.823423 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:47:04 crc kubenswrapper[4800]: E0122 00:47:04.824129 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:47:18 crc kubenswrapper[4800]: I0122 00:47:18.817729 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:47:18 crc kubenswrapper[4800]: E0122 00:47:18.818321 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:47:31 crc kubenswrapper[4800]: I0122 00:47:31.818245 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:47:31 crc kubenswrapper[4800]: E0122 00:47:31.819229 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:47:42 crc kubenswrapper[4800]: I0122 00:47:42.818216 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:47:42 crc kubenswrapper[4800]: E0122 00:47:42.820039 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:47:57 crc kubenswrapper[4800]: I0122 00:47:57.817681 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:47:57 crc kubenswrapper[4800]: E0122 00:47:57.818438 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:48:08 crc kubenswrapper[4800]: I0122 00:48:08.817872 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:48:08 crc kubenswrapper[4800]: E0122 00:48:08.818752 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:48:23 crc kubenswrapper[4800]: I0122 00:48:23.818253 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:48:23 crc kubenswrapper[4800]: E0122 00:48:23.819210 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:48:36 crc kubenswrapper[4800]: I0122 00:48:36.818520 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:48:36 crc kubenswrapper[4800]: E0122 00:48:36.819102 4800 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-mrfxg_openshift-machine-config-operator(8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1)\"" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.974254 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:42 crc kubenswrapper[4800]: E0122 00:48:42.975139 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f899bc4-42c1-4444-aafb-030278227777" containerName="collect-profiles" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.975155 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f899bc4-42c1-4444-aafb-030278227777" containerName="collect-profiles" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.975281 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f899bc4-42c1-4444-aafb-030278227777" containerName="collect-profiles" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.976226 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.992441 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.995223 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.995306 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:42 crc kubenswrapper[4800]: I0122 00:48:42.995410 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d74kd\" (UniqueName: \"kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.096925 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d74kd\" (UniqueName: \"kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.097009 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.097044 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.097506 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.097687 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.142228 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d74kd\" (UniqueName: \"kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd\") pod \"redhat-operators-rmsl2\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.296927 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:43 crc kubenswrapper[4800]: I0122 00:48:43.502610 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:43 crc kubenswrapper[4800]: W0122 00:48:43.511977 4800 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a37752b_f9d1_4c64_9ecf_123dcc5da429.slice/crio-34ef105b4bc921feb8ef379c0145237633bb904b9fc7f1e94c8a1f45eb29eca2 WatchSource:0}: Error finding container 34ef105b4bc921feb8ef379c0145237633bb904b9fc7f1e94c8a1f45eb29eca2: Status 404 returned error can't find the container with id 34ef105b4bc921feb8ef379c0145237633bb904b9fc7f1e94c8a1f45eb29eca2 Jan 22 00:48:44 crc kubenswrapper[4800]: I0122 00:48:44.414815 4800 generic.go:334] "Generic (PLEG): container finished" podID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerID="00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a" exitCode=0 Jan 22 00:48:44 crc kubenswrapper[4800]: I0122 00:48:44.414873 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerDied","Data":"00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a"} Jan 22 00:48:44 crc kubenswrapper[4800]: I0122 00:48:44.415986 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerStarted","Data":"34ef105b4bc921feb8ef379c0145237633bb904b9fc7f1e94c8a1f45eb29eca2"} Jan 22 00:48:45 crc kubenswrapper[4800]: I0122 00:48:45.423440 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerStarted","Data":"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42"} Jan 22 00:48:46 crc kubenswrapper[4800]: I0122 00:48:46.430423 4800 generic.go:334] "Generic (PLEG): container finished" podID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerID="9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42" exitCode=0 Jan 22 00:48:46 crc kubenswrapper[4800]: I0122 00:48:46.430468 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerDied","Data":"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42"} Jan 22 00:48:47 crc kubenswrapper[4800]: I0122 00:48:47.436945 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerStarted","Data":"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04"} Jan 22 00:48:47 crc kubenswrapper[4800]: I0122 00:48:47.453227 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rmsl2" podStartSLOduration=3.047304172 podStartE2EDuration="5.453213775s" podCreationTimestamp="2026-01-22 00:48:42 +0000 UTC" firstStartedPulling="2026-01-22 00:48:44.417083357 +0000 UTC m=+2499.897386385" lastFinishedPulling="2026-01-22 00:48:46.82299295 +0000 UTC m=+2502.303295988" observedRunningTime="2026-01-22 00:48:47.451103048 +0000 UTC m=+2502.931406086" watchObservedRunningTime="2026-01-22 00:48:47.453213775 +0000 UTC m=+2502.933516803" Jan 22 00:48:51 crc kubenswrapper[4800]: I0122 00:48:51.817606 4800 scope.go:117] "RemoveContainer" containerID="84c8d927f2d7e9835c2dea67bab1a73bb33909c97fd2dc166163ba8021050ee2" Jan 22 00:48:52 crc kubenswrapper[4800]: I0122 00:48:52.466511 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" event={"ID":"8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1","Type":"ContainerStarted","Data":"b7125256926d209745f891b64a260e7d38925819555e0e27732c1c9ef0e5ba54"} Jan 22 00:48:53 crc kubenswrapper[4800]: I0122 00:48:53.298048 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:53 crc kubenswrapper[4800]: I0122 00:48:53.298904 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:53 crc kubenswrapper[4800]: I0122 00:48:53.336648 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:53 crc kubenswrapper[4800]: I0122 00:48:53.507991 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:53 crc kubenswrapper[4800]: I0122 00:48:53.564665 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.483064 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rmsl2" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="registry-server" containerID="cri-o://f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04" gracePeriod=2 Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.834461 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.967170 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content\") pod \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.967263 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities\") pod \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.967363 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d74kd\" (UniqueName: \"kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd\") pod \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\" (UID: \"3a37752b-f9d1-4c64-9ecf-123dcc5da429\") " Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.968674 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities" (OuterVolumeSpecName: "utilities") pod "3a37752b-f9d1-4c64-9ecf-123dcc5da429" (UID: "3a37752b-f9d1-4c64-9ecf-123dcc5da429"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.969069 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:48:55 crc kubenswrapper[4800]: I0122 00:48:55.973704 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd" (OuterVolumeSpecName: "kube-api-access-d74kd") pod "3a37752b-f9d1-4c64-9ecf-123dcc5da429" (UID: "3a37752b-f9d1-4c64-9ecf-123dcc5da429"). InnerVolumeSpecName "kube-api-access-d74kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.070670 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d74kd\" (UniqueName: \"kubernetes.io/projected/3a37752b-f9d1-4c64-9ecf-123dcc5da429-kube-api-access-d74kd\") on node \"crc\" DevicePath \"\"" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.490050 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rmsl2" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.490058 4800 generic.go:334] "Generic (PLEG): container finished" podID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerID="f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04" exitCode=0 Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.490078 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerDied","Data":"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04"} Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.490526 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rmsl2" event={"ID":"3a37752b-f9d1-4c64-9ecf-123dcc5da429","Type":"ContainerDied","Data":"34ef105b4bc921feb8ef379c0145237633bb904b9fc7f1e94c8a1f45eb29eca2"} Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.490559 4800 scope.go:117] "RemoveContainer" containerID="f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.525327 4800 scope.go:117] "RemoveContainer" containerID="9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.547159 4800 scope.go:117] "RemoveContainer" containerID="00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.567734 4800 scope.go:117] "RemoveContainer" containerID="f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04" Jan 22 00:48:56 crc kubenswrapper[4800]: E0122 00:48:56.568240 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04\": container with ID starting with f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04 not found: ID does not exist" containerID="f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.568272 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04"} err="failed to get container status \"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04\": rpc error: code = NotFound desc = could not find container \"f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04\": container with ID starting with f1ca80a516f0a8dd9d7663eb2c220b88e07f6ee224eaf3601a72d84b3a770a04 not found: ID does not exist" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.568299 4800 scope.go:117] "RemoveContainer" containerID="9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42" Jan 22 00:48:56 crc kubenswrapper[4800]: E0122 00:48:56.568520 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42\": container with ID starting with 9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42 not found: ID does not exist" containerID="9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.568541 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42"} err="failed to get container status \"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42\": rpc error: code = NotFound desc = could not find container \"9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42\": container with ID starting with 9e13fd794576af35b3029be54ad234569da6061dd8a5567a58e17a7359353b42 not found: ID does not exist" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.568556 4800 scope.go:117] "RemoveContainer" containerID="00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a" Jan 22 00:48:56 crc kubenswrapper[4800]: E0122 00:48:56.568765 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a\": container with ID starting with 00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a not found: ID does not exist" containerID="00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a" Jan 22 00:48:56 crc kubenswrapper[4800]: I0122 00:48:56.568784 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a"} err="failed to get container status \"00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a\": rpc error: code = NotFound desc = could not find container \"00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a\": container with ID starting with 00c137afda91d8bdfea65c499772b10a1760280ec0ed19921a8b70a544286b2a not found: ID does not exist" Jan 22 00:48:57 crc kubenswrapper[4800]: I0122 00:48:57.851940 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a37752b-f9d1-4c64-9ecf-123dcc5da429" (UID: "3a37752b-f9d1-4c64-9ecf-123dcc5da429"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:48:57 crc kubenswrapper[4800]: I0122 00:48:57.932681 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a37752b-f9d1-4c64-9ecf-123dcc5da429-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:48:58 crc kubenswrapper[4800]: I0122 00:48:58.020705 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:58 crc kubenswrapper[4800]: I0122 00:48:58.024850 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rmsl2"] Jan 22 00:48:58 crc kubenswrapper[4800]: I0122 00:48:58.825906 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" path="/var/lib/kubelet/pods/3a37752b-f9d1-4c64-9ecf-123dcc5da429/volumes" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.056046 4800 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:42 crc kubenswrapper[4800]: E0122 00:49:42.057212 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="registry-server" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.057236 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="registry-server" Jan 22 00:49:42 crc kubenswrapper[4800]: E0122 00:49:42.057273 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="extract-utilities" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.057285 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="extract-utilities" Jan 22 00:49:42 crc kubenswrapper[4800]: E0122 00:49:42.057306 4800 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="extract-content" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.057319 4800 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="extract-content" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.057486 4800 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a37752b-f9d1-4c64-9ecf-123dcc5da429" containerName="registry-server" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.058833 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.073628 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.194400 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.194485 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml9vx\" (UniqueName: \"kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.194512 4800 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.295636 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.295699 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml9vx\" (UniqueName: \"kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.295719 4800 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.296125 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.296161 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.315770 4800 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml9vx\" (UniqueName: \"kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx\") pod \"community-operators-vssgq\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.384414 4800 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.675550 4800 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.808332 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerStarted","Data":"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c"} Jan 22 00:49:42 crc kubenswrapper[4800]: I0122 00:49:42.808371 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerStarted","Data":"936b0779e34e6cd2b497abd01a9238678dbf8eb07643280f8429c6ae4fce289e"} Jan 22 00:49:43 crc kubenswrapper[4800]: I0122 00:49:43.816542 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerDied","Data":"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c"} Jan 22 00:49:43 crc kubenswrapper[4800]: I0122 00:49:43.816378 4800 generic.go:334] "Generic (PLEG): container finished" podID="3e334e67-2d37-4ef1-b2b5-608cd56b74c3" containerID="7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c" exitCode=0 Jan 22 00:49:43 crc kubenswrapper[4800]: I0122 00:49:43.820297 4800 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 00:49:44 crc kubenswrapper[4800]: I0122 00:49:44.835984 4800 generic.go:334] "Generic (PLEG): container finished" podID="3e334e67-2d37-4ef1-b2b5-608cd56b74c3" containerID="1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00" exitCode=0 Jan 22 00:49:44 crc kubenswrapper[4800]: I0122 00:49:44.836539 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerDied","Data":"1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00"} Jan 22 00:49:45 crc kubenswrapper[4800]: I0122 00:49:45.854188 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerStarted","Data":"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6"} Jan 22 00:49:45 crc kubenswrapper[4800]: I0122 00:49:45.884425 4800 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vssgq" podStartSLOduration=2.492703756 podStartE2EDuration="3.884409124s" podCreationTimestamp="2026-01-22 00:49:42 +0000 UTC" firstStartedPulling="2026-01-22 00:49:43.819699331 +0000 UTC m=+2559.300002409" lastFinishedPulling="2026-01-22 00:49:45.211404739 +0000 UTC m=+2560.691707777" observedRunningTime="2026-01-22 00:49:45.88273762 +0000 UTC m=+2561.363040648" watchObservedRunningTime="2026-01-22 00:49:45.884409124 +0000 UTC m=+2561.364712172" Jan 22 00:49:46 crc kubenswrapper[4800]: I0122 00:49:46.903547 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wrmd4/must-gather-wcv62"] Jan 22 00:49:46 crc kubenswrapper[4800]: I0122 00:49:46.903780 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wrmd4/must-gather-wcv62" podUID="51656a67-bab8-48ce-bcd1-89209dfa34a0" containerName="gather" containerID="cri-o://e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9" gracePeriod=2 Jan 22 00:49:46 crc kubenswrapper[4800]: I0122 00:49:46.903912 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wrmd4/must-gather-wcv62" podUID="51656a67-bab8-48ce-bcd1-89209dfa34a0" containerName="copy" containerID="cri-o://3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6" gracePeriod=2 Jan 22 00:49:46 crc kubenswrapper[4800]: I0122 00:49:46.913412 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wrmd4/must-gather-wcv62"] Jan 22 00:49:47 crc kubenswrapper[4800]: I0122 00:49:47.866310 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wrmd4_must-gather-wcv62_51656a67-bab8-48ce-bcd1-89209dfa34a0/copy/0.log" Jan 22 00:49:47 crc kubenswrapper[4800]: I0122 00:49:47.866860 4800 generic.go:334] "Generic (PLEG): container finished" podID="51656a67-bab8-48ce-bcd1-89209dfa34a0" containerID="3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6" exitCode=143 Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.182600 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wrmd4_must-gather-wcv62_51656a67-bab8-48ce-bcd1-89209dfa34a0/copy/0.log" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.183278 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wrmd4_must-gather-wcv62_51656a67-bab8-48ce-bcd1-89209dfa34a0/gather/0.log" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.183346 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.290381 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output\") pod \"51656a67-bab8-48ce-bcd1-89209dfa34a0\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.290457 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntx59\" (UniqueName: \"kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59\") pod \"51656a67-bab8-48ce-bcd1-89209dfa34a0\" (UID: \"51656a67-bab8-48ce-bcd1-89209dfa34a0\") " Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.295984 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59" (OuterVolumeSpecName: "kube-api-access-ntx59") pod "51656a67-bab8-48ce-bcd1-89209dfa34a0" (UID: "51656a67-bab8-48ce-bcd1-89209dfa34a0"). InnerVolumeSpecName "kube-api-access-ntx59". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.360131 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "51656a67-bab8-48ce-bcd1-89209dfa34a0" (UID: "51656a67-bab8-48ce-bcd1-89209dfa34a0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.391587 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntx59\" (UniqueName: \"kubernetes.io/projected/51656a67-bab8-48ce-bcd1-89209dfa34a0-kube-api-access-ntx59\") on node \"crc\" DevicePath \"\"" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.391620 4800 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/51656a67-bab8-48ce-bcd1-89209dfa34a0-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.882027 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wrmd4_must-gather-wcv62_51656a67-bab8-48ce-bcd1-89209dfa34a0/copy/0.log" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.882483 4800 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wrmd4_must-gather-wcv62_51656a67-bab8-48ce-bcd1-89209dfa34a0/gather/0.log" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.882525 4800 generic.go:334] "Generic (PLEG): container finished" podID="51656a67-bab8-48ce-bcd1-89209dfa34a0" containerID="e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9" exitCode=137 Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.882573 4800 scope.go:117] "RemoveContainer" containerID="3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.882584 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wrmd4/must-gather-wcv62" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.906466 4800 scope.go:117] "RemoveContainer" containerID="e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.943033 4800 scope.go:117] "RemoveContainer" containerID="3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6" Jan 22 00:49:49 crc kubenswrapper[4800]: E0122 00:49:49.943405 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6\": container with ID starting with 3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6 not found: ID does not exist" containerID="3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.943428 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6"} err="failed to get container status \"3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6\": rpc error: code = NotFound desc = could not find container \"3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6\": container with ID starting with 3ca4173b99a8e6df2ff121c06b7d7a3e2b518dafa9a4411122c5a0a8fe0eb1f6 not found: ID does not exist" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.943447 4800 scope.go:117] "RemoveContainer" containerID="e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9" Jan 22 00:49:49 crc kubenswrapper[4800]: E0122 00:49:49.943958 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9\": container with ID starting with e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9 not found: ID does not exist" containerID="e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9" Jan 22 00:49:49 crc kubenswrapper[4800]: I0122 00:49:49.943983 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9"} err="failed to get container status \"e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9\": rpc error: code = NotFound desc = could not find container \"e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9\": container with ID starting with e31caf0b9896757fa89618e2b243251c0385716a9b436ced2fa3bfeccc094ba9 not found: ID does not exist" Jan 22 00:49:50 crc kubenswrapper[4800]: I0122 00:49:50.826605 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51656a67-bab8-48ce-bcd1-89209dfa34a0" path="/var/lib/kubelet/pods/51656a67-bab8-48ce-bcd1-89209dfa34a0/volumes" Jan 22 00:49:52 crc kubenswrapper[4800]: I0122 00:49:52.384797 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:52 crc kubenswrapper[4800]: I0122 00:49:52.384856 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:52 crc kubenswrapper[4800]: I0122 00:49:52.485061 4800 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:52 crc kubenswrapper[4800]: I0122 00:49:52.936931 4800 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:52 crc kubenswrapper[4800]: I0122 00:49:52.986725 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:54 crc kubenswrapper[4800]: I0122 00:49:54.921340 4800 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vssgq" podUID="3e334e67-2d37-4ef1-b2b5-608cd56b74c3" containerName="registry-server" containerID="cri-o://56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6" gracePeriod=2 Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.327222 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.469556 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content\") pod \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.469603 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities\") pod \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.469641 4800 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml9vx\" (UniqueName: \"kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx\") pod \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\" (UID: \"3e334e67-2d37-4ef1-b2b5-608cd56b74c3\") " Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.470776 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities" (OuterVolumeSpecName: "utilities") pod "3e334e67-2d37-4ef1-b2b5-608cd56b74c3" (UID: "3e334e67-2d37-4ef1-b2b5-608cd56b74c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.471119 4800 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.476031 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx" (OuterVolumeSpecName: "kube-api-access-ml9vx") pod "3e334e67-2d37-4ef1-b2b5-608cd56b74c3" (UID: "3e334e67-2d37-4ef1-b2b5-608cd56b74c3"). InnerVolumeSpecName "kube-api-access-ml9vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.541330 4800 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e334e67-2d37-4ef1-b2b5-608cd56b74c3" (UID: "3e334e67-2d37-4ef1-b2b5-608cd56b74c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.572461 4800 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.572487 4800 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml9vx\" (UniqueName: \"kubernetes.io/projected/3e334e67-2d37-4ef1-b2b5-608cd56b74c3-kube-api-access-ml9vx\") on node \"crc\" DevicePath \"\"" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.928060 4800 generic.go:334] "Generic (PLEG): container finished" podID="3e334e67-2d37-4ef1-b2b5-608cd56b74c3" containerID="56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6" exitCode=0 Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.928130 4800 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vssgq" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.928136 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerDied","Data":"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6"} Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.928468 4800 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vssgq" event={"ID":"3e334e67-2d37-4ef1-b2b5-608cd56b74c3","Type":"ContainerDied","Data":"936b0779e34e6cd2b497abd01a9238678dbf8eb07643280f8429c6ae4fce289e"} Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.928489 4800 scope.go:117] "RemoveContainer" containerID="56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.956118 4800 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.956284 4800 scope.go:117] "RemoveContainer" containerID="1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00" Jan 22 00:49:55 crc kubenswrapper[4800]: I0122 00:49:55.964234 4800 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vssgq"] Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.001751 4800 scope.go:117] "RemoveContainer" containerID="7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.034423 4800 scope.go:117] "RemoveContainer" containerID="56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6" Jan 22 00:49:56 crc kubenswrapper[4800]: E0122 00:49:56.034996 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6\": container with ID starting with 56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6 not found: ID does not exist" containerID="56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.035032 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6"} err="failed to get container status \"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6\": rpc error: code = NotFound desc = could not find container \"56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6\": container with ID starting with 56e2c654c7094dc284b564b692ef0aa4af4993d459f9bf2773659e188bc204d6 not found: ID does not exist" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.035061 4800 scope.go:117] "RemoveContainer" containerID="1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00" Jan 22 00:49:56 crc kubenswrapper[4800]: E0122 00:49:56.035326 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00\": container with ID starting with 1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00 not found: ID does not exist" containerID="1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.035354 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00"} err="failed to get container status \"1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00\": rpc error: code = NotFound desc = could not find container \"1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00\": container with ID starting with 1994494b1946b1d2881bffaf3f1bdd19b7bd6b5c39fe630c0693b9f7a0034b00 not found: ID does not exist" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.035400 4800 scope.go:117] "RemoveContainer" containerID="7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c" Jan 22 00:49:56 crc kubenswrapper[4800]: E0122 00:49:56.035672 4800 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c\": container with ID starting with 7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c not found: ID does not exist" containerID="7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.035710 4800 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c"} err="failed to get container status \"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c\": rpc error: code = NotFound desc = could not find container \"7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c\": container with ID starting with 7d73f874af1d03038bd2f30cb4c82154057b0efde160799f0e22435ff2be524c not found: ID does not exist" Jan 22 00:49:56 crc kubenswrapper[4800]: I0122 00:49:56.827184 4800 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e334e67-2d37-4ef1-b2b5-608cd56b74c3" path="/var/lib/kubelet/pods/3e334e67-2d37-4ef1-b2b5-608cd56b74c3/volumes" Jan 22 00:51:19 crc kubenswrapper[4800]: I0122 00:51:19.452074 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:51:19 crc kubenswrapper[4800]: I0122 00:51:19.452644 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 00:51:49 crc kubenswrapper[4800]: I0122 00:51:49.452304 4800 patch_prober.go:28] interesting pod/machine-config-daemon-mrfxg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 00:51:49 crc kubenswrapper[4800]: I0122 00:51:49.452861 4800 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-mrfxg" podUID="8e02d6ab-3bcb-4fb2-a918-050e2b87ecd1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134272266024455 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134272267017373 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134264461016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134264461015463 5ustar corecore

ָ߲s\7dH{6gόjlr֨p~"tzyN4pQhEpg]}0K/UO?o f^V<Ѡo_A=YVi·g8NǓޟO_x=_$|nӼ/^ׯƾӼWç/Б>CM~?h%!7_)g< F_z15 \s2g~&M=LJ/bft1~׈_1?\&kO@"s~`ݻSo0䓄-_QEb8~|js- Wc>ͷ&cExda*^g,}YGv6fi4$3;+O>x7)"fDҀWdS~-U׹S# L+KWHV ~fv)E>i/ҙ~inrC~b)O2nɆٺ:|x%} 5gs,]?‰1ǖ?,^8cr|}[|ɕsk}m}ߝ }}:rk7%"(&_`<DB$w*ϊK˦lTlTqM؏%KA7JY\- yrɖK\^ %MEFJ|oJ-$<2FX`i%d<+./ MTJ8 O?-GDF[^məe1|+ 1?fg+k9Wؖg{S rx~بX}eoO'Jw-(Wp($l({j.6&_2mJY_ b.4肦 B}5g \یpD V?9:3Q:k;x"5ϑpymn6WJh?˨ſq0م x,vdVd Ƿe=d|k?2)?dt'8x*'LA`FLH%[5)VmOUB`OZّK0\ޭT*%g .Q9DT5D$abbXmL1P88GL:gN:bM{b)9y<,#p rP{b?7wȞ~?Óܧc(|L'68w#K A]F4FCaP7P oKkP?:ſ9:x0tP.?Fc\skr͙_9+.̿f&Qذ8I៑&I0J0KM4EViǻE 3mPNʓ{<ԨPO_=]˲թ^mv]vyԉe>K37~yeŞ: Ts-Cj]4naos7nE0 ( .uCr1fJHթ2*Ʊ1L#RV+*IʠbE,TzDEY kn@GXuGOx6Ĵ)x6Ҭ p|pʬD<ť'U(lK/c5m827t'@]jFIVH2ɽP 2ɖIL ILeeu]u HQqg2mtR#cJc[&$3Y1u \ٕ])1p+!u*2P&c8K4V !2M]b'?{ɭdFTx7A|b>Γ &ׂF;{=#i.={#_KUn*JARA 2٭*ѼKfW梹}#z_m&ǽFRJmhf^rܡ0 hQtQrv%G[&EAAwW֓*tq]TN4Q;6ъH;QE V^zC,0̬A+oO=:&kA dF PΪDdAjoLV$¿sP;|]P;>e3/ʵ뢸]?[>594l haI{Yto^t.v˳2?:zYJiOGD+PDAQtXYjd ]Nc`ht"[AX"{puaXzm;&lP?Vv6;-$a~#v3."Y@fvaꁝib=Q䔳No{ɸ3?rv W@+j&[t@W՚I(bw" -zRհg&m1>(Lv ;|5צw/ry(.uI֡F D^01/9\%ziq @7O$Lrgv~]tQ+VP|%YqDT.ߑz:"620d-&AnU*@6"T#v 6=T%Ł4^+潉sEZM0.hrL(VyJtXFv0|GCt?W[?4C"A#s!BHṔcN NYc8|uzR9> . mNρ%Ne&xN:<ΊF]oRux˯Bp}Tsn·ArPHL-Wz U%w!|)4R˦Rȁ;=4c+L=jkwu7rzA iucQ 5 (anP^}k@*sOp:`{.&hφQJpƳ@҉ :w?6IR F*ӱqyzB6jӵO9a(ʻ6p`H:v4`l|k΂1԰т\18/΅RΨ2~ R^7:DqfcK3ζ8S׫3Eo5 c3h:^v+oٗq;;[xi)<<食fyna2Nvz]Jd3^v%8޷_f3Wf\]p{#:ӓke7@ʜ7M1w%-D(nk;\fR9@SPD~gdѰDyFPIrwʦ:(llb OD~`l @fT{FzXbNTP9EΗQTEZx)u/C^FEC(aQ)xZiߘdZhضff$g} ѣ~$)5%U3EQZSQPDByw&yQ2QVaOM⨕qC.WL~oqvfEPFl2H>_^|/y-FI-̽ 2jGI .lbNkk .aMd>t|O3t7@G?vg p=)韼s{2kGT {{䝠i{d=0&[-]Jv}$l?*=ɂdM.q&JQ *P{Gfm+y[|i a_у|G :Gy徻AmO{Bd[փ<IRcd#, ͟VlLWve@ނ}ROw&vr䕝]r>.Jggivu~[-~Ι|ړw'__#yC^'wO;_>;;#9Kl18c'w':~ʛo;Np~^do^yd#{ێs6+$C{YhPaCJA`Me)uy'xI273lTb$(F)/ &&^p uVAjU\Yjji/?CY.jd6/uh\Xk2b /IA887v "Zˇ]'Xm{@o庄/?*Ҹ]Gv;j4[c"n.4*dSd[<%'[dvewV+NF fhR 99tZOeS*- mxUț4Q7A/ӎ笰5n^  aJZ onY5v;G$A RN}HRYy Pv6V){ PI JW~`Sή_#6X` ,yѮm^ΗpŅ>3\ ;떻TM~--lQO558UiaC`R=n4( eoBi Q2?1zwLÍ0d6>0yf)Yc1`pASD#x,BHTḾs3X( >lH:(OX$\A؈h`lL8mmlۨCҘN*Pc-w,aZʿ?t-_ֺn>߯9&dmVN Pw5r^WZcQU֞|ǻ{{U$EbVpK$F$FcdzD @ 3"΀-,H%g2Zlڢ3ZIh4+54$JpTE=bYqEC+=D«FX:PF1]$u:>Ө[aIk|$"Vǘ1Ơ&Elnei?IWz$B[ɷWrtݭ3USs,wuK=GֽFއ8Z,F?(?9Nu'~8TCMf 46NcJ㸑2Зȼ"ƈ쇗`SVkG#"+Όԣa7qtpY+ Q5r|E;@hI IO=AYtT6 `!n<*"*| P#Can%j}.8 /4]^cJ[:geL>p,UtΜ̮/I_^^hÛcߝvݏ]>/h˓c ۹ZI9|{]"h)Gh9&^גӅ$i` ,M(iGoVfv|U.U^')HIEJ"-I*ɽN.RjuRZ]TSUrKYLj=u-,3GZn׫M &5~"H%R,٬ރDu؋&Ǝw\Wg 9\W<@ÁXKiX-j-*y㹯{_xW0`>$7qNEs0/|aL>>݊U-7mFHAf'?xq=1i}:QMX/UoXi.iX=)@A{=z@J"kO+dYrC$Bb(p}MM7 ŁNc*ge}пξ)0$MmMJKZF5"iE1UQyyGz_^G$ Lc" :$.2wW:wwal[!v^`R*zc_ւM3Y*Z69Ls=w30ECe %mu[ޠs ,ZToI4)9=8,u\+5ֶuE Z x%L*_<~xeegSSicK{e `GSAyE% $2㐤"jE!G8Zjq >Di6^Z:\!+Q l%eql-;#w1cKVƕ<}׮wy`^G7ˣa=: ,`x&w2wDbnUC.*MhJĨEu~'bąFw(wĨ%bh%P"F ĮQYNHQI]PRVRuTY<}mɻ!`PMn5JY"EQ;MPnG6x'z)u,Lx k(=76&`O;y`rd;g SQ$wp"&tV[U8s,U!дk6m\gj7S_L)cTy7`_<19j*F6"j-_/x>S̉\}b9ǫ3&X7͒Tb|&`CK6)yƚgRxsCƜ]V+!s%R?=tQNd /GsOa|r #ofe7?"[w%;1esJ{cc躼@2f3k B`O,_~V#jwnϭ2r6k5ogߝwՇ'WGg [؞mǹdH!++&9Y\^f'~l367E6%93kRwD2|V^17|`FĀG݌idu`1S)xljD l|p(-JtŅ^1.:=#sw9Ϩ%31_QdIwLZ%[\-/=ec()ZB#z;E؃yM8gNLdӇm:(cO{2蹕{z63y﶑άSkؐE"lhUܢ5xha%ph5]䣕WViE4eYYQYV ,ɇlD71 U'wDݳ݌|_F'Tb14Gf݇Ӵdݛ eH nN?Bc8Lhn u&4{+sǶy^KyFW&h>T=Z'C$RAҽQob*'.#Ƿ#1֜hoL cA'#(XoRg "֔ʉ QH$4N3epVCVEŖY!My(ЭupMwpz!̺vVgMSo>i0JlAzOe ɕ#4:A}2w9]t<$fH<9"&Uj& [* š 0(6YjM,NEVpJ8T^gS07ؤmyL-ȥ%?F3/t )cuH8EB8BxTőM ~|3hBDh낎0A(k:Z1= 8m@{^"JIimH ûYS\_}K/vpjM94 "ʐzdaBaJ,  pߢ, * e&@YdR1LO w ӽO.\1_Ide4]$Ɓ63ZD( }0T[ @,]FP <h:%n9kCT asg8v* E   o9X`Ϧ*ZZ¤0k# . 6QkqK`8պi7k! D>$oR N !Lq`GLD(Ɣ̚"]9h2JuYL#Ib6vtN R! Z%lJ^Z nA L -F88<(>R a`'.8~"W-ZX<*q,guAM Lʠdy3'@ZRuY. fM9haq5.N8l`ezX{ɕ@_Sn,@   8bDɍ Էed# –@G*!̚W:K`7}2Y\ΖWgT&,S,hIJ¥/Η)%_oSsåՉRWǿj"ջw/7㳣J߱y"yz?.DV}´R >Ycp`Ζau< ~]{wWh~yfv#=QJA񣰪@yn$ "jkk {BdO~TGۉo7yu[]\-N͙9 k]02'U}~e=msЋq4nusW4i8] 7[}vH$@UuD!$REvI^!F[A20 5@r[8 5TQW _XDX ,(m~`BB9wY#pSV IDvU Y LB $k c..CrU*W\'n.+Q&F&RP^1I'KNWc8]j]'1#^'\L1,;0cy]*0>'ߩz驻T@t<zGzW7jL`=f5&.՘b P$AVϗTQՠr ʚ .f|ldԌCjF|K!5Pnb*fl(VEOTݜ9* ITmOB:&U)]\߬n TO^?gfthVjYPWM J5Q]o_DP]'Qx}'ktny~zX4f$AQYan=lNmmM]%MzU<~_~Е58J O(M7ptQ!b˵%=oNI.fˋwہԅyaV^<`N%%%seykN=F| ^":&>|'tB]]OF[ Jz .{B3j A.;u oE}'uRInt'mc4p&#*b' ]Ɣ'+9cn+ _K*Un.OF޷9x_${obW 7SHR.盗}{=i/E'q\yre|}aڠDT` |3k)Acv*V/cT~IҼ)(Oޫ HrWРL[KʟɅa=!g76T\ d/&[u ntM2D](n_m^w3QтDMU$DplK}*JuD3%Œz5=RR4;ޑ]^V_DЎ6?YDmy2=x%$?1RbF`ڲ|-1ңmw>CR,R FGH*fLwJ`I/ϓ^<ũ̝D&ǩr021&᠂őjt2,U`m(k%5`hGqB\ݒׅdCt`<'S L?1tn_^f !+x%T{$D0I !H*AZ3bl_3, Z=Ml4?e`(:ᔽߌIc?H 1*ѡ*)\G<%Q1CCaVcQ;-l.rxp3:Cd:>jQ%D`fi1BSH5pJ>3^bSTXwEQ*Ͼ0t2Hz۳߫/ThGO.~%v_yQD?Ǐڍ./rQ]Oއ>f(iҕ*ٝeZyu>9G|xo$BiZezo=hf%򸟏ۜ{At2!wHsNj5}akٙPRCLFd mc20mV0^,{]Op4#? i0;Їٕ^`f*:fJ"`jG4FS6hi0ZJoE'~Z䗻\r p;^xBdĩ1Q1)`. J1G% ̉VǩbJL 7J#OG޳h"q'Z 5e=LT(h$NüIX 5" Rpt)lM=sbM*mՠkn&7B݄|;T)s9_TҶn{=S%m|cma`lO$jsn%JJ66!fs/}59L`DDTgg\ }:́`4fgOTbCsٛISL)iE"Qp@RY ߊ\(r~fq{RS ;-ddƇ^oktD_Iz_O_}4Wf.,10 #>m8c[;S|>.I*)O '?.[DsNKF`6 |;n<7___^}0xs]5ޤ7O@u~6`r.)0û߱S7i6.:GC^!RcJ'&T MkP@gV $&@լ]^Yߛa/O{ 뢑TTJ{C20',ޅ-~j22a_a `86OCun&H:F0a%zaIRF'r?|L>^˓Ezuxj~&La65?s1|x0Ow;~_p7⒭\(U%oxϤ~P%dɵRm ۹FH?%X*> ,{Md[>z)C(Kq#N< A>W93 Z\0?e@jj(WӣЈ.V ݕT2 ȗ㷅6=.Gmob V  Tx읙 vMu~ >վ3t  ^3GPfz~&6a夞??`CNP ώ"~~[Ik36kQI.j mt AQQ4nNpa=(1A^$0cO86ϣ֞xs(dH,BvExd&y+'"JbE[m(LESJ86E)iMyͪhqQ'1fhlpNЀ!Y*YdRrcc(؃nF8oT+Ag2<j/u]T{}3$X(NVX)Ѩh4E=LKչ,t a<ƾhHMx7t!ɂ)b`VW4͆`| >+gpaPͻ'79lQQr>n1hԥ) vaV,A__3'trQZ~W5.Jlb#C 63DPԈ? KQ,+"O \G:I]GvLw8Fp,"Ё(1pC)xN0R/3je΍k#EЍe:m.C])ۑZTp#E{>RTH1(xwhGjnmz"Xf3N:k1YTxb>Y$RS̜nm.%(oCsYH3۩h<_П2/ivI֢l. l[#i\ufR ģH]>j<*(#ǣsGn暲ee ,d[ee p粏벥V,˱זqLY?Ȇ H TRj$u[.m(1^K̕n^6 QR^t|Uoz= }CzK{;x>vKX»wG Aw{?_f6^9Moq:H0+ 9Ӑ8ebN1Si:U*Ǜt7/N+~ZhO\ \ h3W,Ɔuc}#Z\ Nnuĩ8uukBCƴs &mnTQ6X#3e*f0-hUք % ;@˜0\܆I _g8uޡIj<@v1GW Vf%$wq9O 5`(ޫNLLx37 nzaa-B"-;܍AoAκ0woa4![.d'm:+Z r*vļMRdr7on+f}fb10[( o'H-]E۟^,ѐͧCw"X.méY(YEEj'-H݃zCJR#u R^N )"5XUL@$$20t^ZT|r<.C9ؠIG]W `|=34ǣ  =7hzh*Ydݻ݇t1q|=)IqCޘikj˕Q.Y\_AMm64< 7ǃU-o~e'&.i(tZҪ!y5cgl`E*0xFn_Z;]݅qoᝆޣ׭ 3ڻt:iq]:F# {Rvޟ-"[D7B^+xL9qFv4# b:dJ.3ɣ= Tymǽ]z3Y]^;{Q fJ&>EBW!mXcʓκ)ٿ?Q_ÐRsl1,_?8? 0dϒ?f˯ϛ?_@]yk~gG2sW;dߏO<ȥcʨ˟5ŝI7;6X/ގ|?n!4a5;Xz >̳X\gg M_.ƖZKCN HM861N p1H $ƽ 5.%3;Y,$sÝE h$ ^ ϜgltB>`mXހ=r`:ߎa#RqBQ0?M#|1>t15l.l~Ma l~leiTe( ,SC0s f#iN7ӭFp_ʲRHq_}R"8KY6]ʲ }aeliڋAqmlLq2/TQ 6Lc/wF,o|`!cR *p)( '@rQV>Z  ^|:HXVvGoή";O)\+ D{I I ҏwQtI Ap7#I@FS)@Z<mcRԶ Zč)Q{EsɩJoA"W8Np%E 1ܻrqٔavT+NԈPwI^A!ZjD l:^h jSxU`Y>5S9X':с6]sOFM,Cd`HߋdĈR )F`>[D݇q2nv ;S+4;%vAUw8D x8hqP i;w#IF!:NSG=U <<$8Ck5IA˴5Q>83JxouAT)Hcx/!WlWl|$Np֟TUز/wTX)2~WZUC35zաiա)SvV*'vRa"ufqSc/7dRr, u{odӎԛN^ENJ%ζ MŭI~wߤvs9 ^ #KT{?W\8~+<1³iǼTc݈YN6@rqa^4!S(bo- !BJd=vCNbW[4L|x$04PiPb"Qa_a&ELi\ԅ^ue{{`(.8_]KW=GrpɊ|.+Sԫ.xz攔֣kő;b %űm?0q}t-9!)ZF2.ptkE}Q%m4λ,%,O